var/home/core/zuul-output/0000755000175000017500000000000015114517554014535 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114534334015474 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005332027615114534325017710 0ustar rootrootDec 05 09:06:40 crc systemd[1]: Starting Kubernetes Kubelet... Dec 05 09:06:40 crc restorecon[4652]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:40 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:41 crc restorecon[4652]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 05 09:06:41 crc restorecon[4652]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 05 09:06:41 crc kubenswrapper[4815]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 09:06:41 crc kubenswrapper[4815]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 05 09:06:41 crc kubenswrapper[4815]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 09:06:41 crc kubenswrapper[4815]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 09:06:41 crc kubenswrapper[4815]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 05 09:06:41 crc kubenswrapper[4815]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.243442 4815 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246842 4815 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246863 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246868 4815 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246872 4815 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246875 4815 feature_gate.go:330] unrecognized feature gate: Example Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246880 4815 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246885 4815 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246890 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246894 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246898 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246902 4815 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246907 4815 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246911 4815 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246914 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246918 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246921 4815 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246924 4815 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246928 4815 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246932 4815 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246935 4815 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246939 4815 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246943 4815 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246948 4815 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246952 4815 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246956 4815 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246960 4815 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246963 4815 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246967 4815 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246971 4815 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246974 4815 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246978 4815 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246981 4815 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246985 4815 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246988 4815 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246992 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246995 4815 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.246999 4815 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247002 4815 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247006 4815 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247009 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247013 4815 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247016 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247020 4815 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247023 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247028 4815 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247031 4815 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247035 4815 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247039 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247042 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247045 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247049 4815 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247052 4815 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247057 4815 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247061 4815 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247065 4815 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247069 4815 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247072 4815 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247076 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247080 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247083 4815 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247086 4815 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247090 4815 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247093 4815 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247097 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247101 4815 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247104 4815 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247108 4815 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247111 4815 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247114 4815 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247117 4815 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.247121 4815 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247188 4815 flags.go:64] FLAG: --address="0.0.0.0" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247197 4815 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247205 4815 flags.go:64] FLAG: --anonymous-auth="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247211 4815 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247219 4815 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247223 4815 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247228 4815 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247233 4815 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247238 4815 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247242 4815 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247246 4815 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247250 4815 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247254 4815 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247258 4815 flags.go:64] FLAG: --cgroup-root="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247262 4815 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247266 4815 flags.go:64] FLAG: --client-ca-file="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247270 4815 flags.go:64] FLAG: --cloud-config="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247274 4815 flags.go:64] FLAG: --cloud-provider="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247278 4815 flags.go:64] FLAG: --cluster-dns="[]" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247283 4815 flags.go:64] FLAG: --cluster-domain="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247287 4815 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247291 4815 flags.go:64] FLAG: --config-dir="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247295 4815 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247299 4815 flags.go:64] FLAG: --container-log-max-files="5" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247305 4815 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247309 4815 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247313 4815 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247317 4815 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247321 4815 flags.go:64] FLAG: --contention-profiling="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247325 4815 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247329 4815 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247334 4815 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247339 4815 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247346 4815 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247351 4815 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247356 4815 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247363 4815 flags.go:64] FLAG: --enable-load-reader="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247368 4815 flags.go:64] FLAG: --enable-server="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247373 4815 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247378 4815 flags.go:64] FLAG: --event-burst="100" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247382 4815 flags.go:64] FLAG: --event-qps="50" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247386 4815 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247390 4815 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247394 4815 flags.go:64] FLAG: --eviction-hard="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247400 4815 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247404 4815 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247408 4815 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247412 4815 flags.go:64] FLAG: --eviction-soft="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247416 4815 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247420 4815 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247423 4815 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247428 4815 flags.go:64] FLAG: --experimental-mounter-path="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247432 4815 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247436 4815 flags.go:64] FLAG: --fail-swap-on="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247440 4815 flags.go:64] FLAG: --feature-gates="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247445 4815 flags.go:64] FLAG: --file-check-frequency="20s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247450 4815 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247454 4815 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247458 4815 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247463 4815 flags.go:64] FLAG: --healthz-port="10248" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247467 4815 flags.go:64] FLAG: --help="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247471 4815 flags.go:64] FLAG: --hostname-override="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247475 4815 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247479 4815 flags.go:64] FLAG: --http-check-frequency="20s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247483 4815 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247500 4815 flags.go:64] FLAG: --image-credential-provider-config="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247504 4815 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247509 4815 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247513 4815 flags.go:64] FLAG: --image-service-endpoint="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247517 4815 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247521 4815 flags.go:64] FLAG: --kube-api-burst="100" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247526 4815 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247532 4815 flags.go:64] FLAG: --kube-api-qps="50" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247537 4815 flags.go:64] FLAG: --kube-reserved="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247542 4815 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247547 4815 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247552 4815 flags.go:64] FLAG: --kubelet-cgroups="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247557 4815 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247562 4815 flags.go:64] FLAG: --lock-file="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247566 4815 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247571 4815 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247576 4815 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247584 4815 flags.go:64] FLAG: --log-json-split-stream="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247588 4815 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247592 4815 flags.go:64] FLAG: --log-text-split-stream="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247597 4815 flags.go:64] FLAG: --logging-format="text" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247601 4815 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247605 4815 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247609 4815 flags.go:64] FLAG: --manifest-url="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247614 4815 flags.go:64] FLAG: --manifest-url-header="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247620 4815 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247624 4815 flags.go:64] FLAG: --max-open-files="1000000" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247629 4815 flags.go:64] FLAG: --max-pods="110" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247633 4815 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247637 4815 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247641 4815 flags.go:64] FLAG: --memory-manager-policy="None" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247645 4815 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247649 4815 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247654 4815 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247658 4815 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247667 4815 flags.go:64] FLAG: --node-status-max-images="50" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247672 4815 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247676 4815 flags.go:64] FLAG: --oom-score-adj="-999" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247680 4815 flags.go:64] FLAG: --pod-cidr="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247686 4815 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247694 4815 flags.go:64] FLAG: --pod-manifest-path="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247698 4815 flags.go:64] FLAG: --pod-max-pids="-1" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247703 4815 flags.go:64] FLAG: --pods-per-core="0" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247710 4815 flags.go:64] FLAG: --port="10250" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247714 4815 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247720 4815 flags.go:64] FLAG: --provider-id="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247725 4815 flags.go:64] FLAG: --qos-reserved="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247729 4815 flags.go:64] FLAG: --read-only-port="10255" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247735 4815 flags.go:64] FLAG: --register-node="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247740 4815 flags.go:64] FLAG: --register-schedulable="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247745 4815 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247752 4815 flags.go:64] FLAG: --registry-burst="10" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247756 4815 flags.go:64] FLAG: --registry-qps="5" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247759 4815 flags.go:64] FLAG: --reserved-cpus="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247764 4815 flags.go:64] FLAG: --reserved-memory="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247768 4815 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247772 4815 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247777 4815 flags.go:64] FLAG: --rotate-certificates="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247781 4815 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247785 4815 flags.go:64] FLAG: --runonce="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247789 4815 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247793 4815 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247797 4815 flags.go:64] FLAG: --seccomp-default="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247802 4815 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247806 4815 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247811 4815 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247815 4815 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247819 4815 flags.go:64] FLAG: --storage-driver-password="root" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247823 4815 flags.go:64] FLAG: --storage-driver-secure="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247827 4815 flags.go:64] FLAG: --storage-driver-table="stats" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247831 4815 flags.go:64] FLAG: --storage-driver-user="root" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247835 4815 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247840 4815 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247844 4815 flags.go:64] FLAG: --system-cgroups="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247848 4815 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247854 4815 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247858 4815 flags.go:64] FLAG: --tls-cert-file="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247862 4815 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247869 4815 flags.go:64] FLAG: --tls-min-version="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247873 4815 flags.go:64] FLAG: --tls-private-key-file="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247877 4815 flags.go:64] FLAG: --topology-manager-policy="none" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247881 4815 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247886 4815 flags.go:64] FLAG: --topology-manager-scope="container" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247891 4815 flags.go:64] FLAG: --v="2" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247897 4815 flags.go:64] FLAG: --version="false" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247904 4815 flags.go:64] FLAG: --vmodule="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247914 4815 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.247919 4815 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248030 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248037 4815 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248047 4815 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248051 4815 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248055 4815 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248059 4815 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248065 4815 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248070 4815 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248076 4815 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248082 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248087 4815 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248091 4815 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248096 4815 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248102 4815 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248107 4815 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248113 4815 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248117 4815 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248122 4815 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248126 4815 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248131 4815 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248135 4815 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248139 4815 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248143 4815 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248148 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248152 4815 feature_gate.go:330] unrecognized feature gate: Example Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248156 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248161 4815 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248166 4815 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248170 4815 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248175 4815 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248179 4815 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248183 4815 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248187 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248192 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248198 4815 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248204 4815 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248209 4815 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248215 4815 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248219 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248224 4815 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248229 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248234 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248238 4815 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248243 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248247 4815 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248252 4815 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248256 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248260 4815 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248265 4815 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248269 4815 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248273 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248277 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248282 4815 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248285 4815 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248290 4815 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248294 4815 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248297 4815 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248301 4815 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248305 4815 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248310 4815 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248314 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248318 4815 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248323 4815 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248328 4815 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248333 4815 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248338 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248345 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248350 4815 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248354 4815 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248358 4815 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.248362 4815 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.248369 4815 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.262435 4815 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.262534 4815 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262707 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262731 4815 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262741 4815 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262750 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262761 4815 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262773 4815 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262783 4815 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262792 4815 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262800 4815 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262809 4815 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262817 4815 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262825 4815 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262834 4815 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262842 4815 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262851 4815 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262859 4815 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262866 4815 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262875 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262883 4815 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262894 4815 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262904 4815 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262914 4815 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262922 4815 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262932 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262942 4815 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262954 4815 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262962 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262972 4815 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262980 4815 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262989 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.262997 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263007 4815 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263014 4815 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263022 4815 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263030 4815 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263038 4815 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263045 4815 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263053 4815 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263061 4815 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263069 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263077 4815 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263084 4815 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263092 4815 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263099 4815 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263108 4815 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263115 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263123 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263130 4815 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263140 4815 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263150 4815 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263160 4815 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263170 4815 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263181 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263190 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263204 4815 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263216 4815 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263227 4815 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263237 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263246 4815 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263255 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263264 4815 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263272 4815 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263280 4815 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263289 4815 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263298 4815 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263306 4815 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263314 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263321 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263329 4815 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263336 4815 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263344 4815 feature_gate.go:330] unrecognized feature gate: Example Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.263357 4815 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263614 4815 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263628 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263638 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263648 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263660 4815 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263670 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263678 4815 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263686 4815 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263695 4815 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263703 4815 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263711 4815 feature_gate.go:330] unrecognized feature gate: Example Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263719 4815 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263727 4815 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263736 4815 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263745 4815 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263753 4815 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263761 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263771 4815 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263781 4815 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263789 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263797 4815 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263804 4815 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263812 4815 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263821 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263828 4815 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263836 4815 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263843 4815 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263851 4815 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263859 4815 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263867 4815 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263874 4815 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263881 4815 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263889 4815 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263896 4815 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263905 4815 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263915 4815 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263927 4815 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263936 4815 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263944 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263954 4815 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263962 4815 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263970 4815 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263979 4815 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263986 4815 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.263994 4815 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264002 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264009 4815 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264017 4815 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264025 4815 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264033 4815 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264041 4815 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264048 4815 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264056 4815 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264063 4815 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264071 4815 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264080 4815 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264089 4815 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264098 4815 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264107 4815 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264115 4815 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264122 4815 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264130 4815 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264137 4815 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264145 4815 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264153 4815 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264161 4815 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264169 4815 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264177 4815 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264185 4815 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264192 4815 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.264200 4815 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.264212 4815 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.264796 4815 server.go:940] "Client rotation is on, will bootstrap in background" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.269109 4815 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.269251 4815 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.270082 4815 server.go:997] "Starting client certificate rotation" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.270130 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.270371 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-11-27 04:15:22.764593685 +0000 UTC Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.270618 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.278258 4815 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.280381 4815 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.67:6443: connect: connection refused" logger="UnhandledError" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.283210 4815 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.290342 4815 log.go:25] "Validated CRI v1 runtime API" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.307216 4815 log.go:25] "Validated CRI v1 image API" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.309520 4815 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.312944 4815 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-05-09-00-20-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.313000 4815 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.338566 4815 manager.go:217] Machine: {Timestamp:2025-12-05 09:06:41.336878643 +0000 UTC m=+0.215485560 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2799998 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:25928f9c-29aa-4c55-87c5-643012e42318 BootID:e9726a9b-025c-49da-b818-99a967ee21f0 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:c4:1b:e6 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:c4:1b:e6 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:11:2a:f8 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:9f:a0:97 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:92:a1:7b Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:73:b8:ed Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:69:3d:fd Speed:-1 Mtu:1496} {Name:eth10 MacAddress:36:ef:00:34:cd:0f Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:da:55:79:3d:c5:c3 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.338905 4815 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.339163 4815 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.339711 4815 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.340074 4815 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.340131 4815 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.340447 4815 topology_manager.go:138] "Creating topology manager with none policy" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.340465 4815 container_manager_linux.go:303] "Creating device plugin manager" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.340811 4815 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.340864 4815 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.341232 4815 state_mem.go:36] "Initialized new in-memory state store" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.341363 4815 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.342324 4815 kubelet.go:418] "Attempting to sync node with API server" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.342369 4815 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.342407 4815 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.342428 4815 kubelet.go:324] "Adding apiserver pod source" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.342451 4815 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.344578 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.67:6443: connect: connection refused Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.344690 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.67:6443: connect: connection refused" logger="UnhandledError" Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.344636 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.67:6443: connect: connection refused Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.344762 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.67:6443: connect: connection refused" logger="UnhandledError" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.344877 4815 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.345300 4815 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.346465 4815 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347350 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347393 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347409 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347424 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347455 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347470 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347485 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347537 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347557 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347572 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347613 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347627 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.347897 4815 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.348591 4815 server.go:1280] "Started kubelet" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.348840 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.67:6443: connect: connection refused Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.348983 4815 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.348985 4815 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.349773 4815 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 05 09:06:41 crc systemd[1]: Started Kubernetes Kubelet. Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.350787 4815 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.350836 4815 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.351434 4815 server.go:460] "Adding debug handlers to kubelet server" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.354479 4815 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.354577 4815 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.354746 4815 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.354767 4815 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.129.56.67:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187e467bc35e7703 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 09:06:41.348482819 +0000 UTC m=+0.227089696,LastTimestamp:2025-12-05 09:06:41.348482819 +0000 UTC m=+0.227089696,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.355449 4815 factory.go:55] Registering systemd factory Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.355469 4815 factory.go:221] Registration of the systemd container factory successfully Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.355654 4815 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 22:37:15.834369975 +0000 UTC Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.355710 4815 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 901h30m34.478664602s for next certificate rotation Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.356093 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.356172 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" interval="200ms" Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.356474 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.67:6443: connect: connection refused Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.356596 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.67:6443: connect: connection refused" logger="UnhandledError" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.358865 4815 factory.go:153] Registering CRI-O factory Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.358902 4815 factory.go:221] Registration of the crio container factory successfully Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.358979 4815 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.359016 4815 factory.go:103] Registering Raw factory Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.359037 4815 manager.go:1196] Started watching for new ooms in manager Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.359985 4815 manager.go:319] Starting recovery of all containers Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375232 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375279 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375294 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375307 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375320 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375333 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375344 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375375 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375390 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375402 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375430 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375442 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375453 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375465 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375479 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375511 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375523 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375536 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375548 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375559 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375571 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375583 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375596 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375608 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375619 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375630 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375644 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375656 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375667 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375678 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375690 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375702 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375713 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375725 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375737 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375749 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375762 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375773 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375785 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375797 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375808 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375819 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375831 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375843 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375855 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375866 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375879 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375891 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375901 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375912 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375924 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375937 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375954 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375968 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375980 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.375991 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376005 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376016 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376027 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376069 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376085 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376097 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376108 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376119 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376131 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376142 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376153 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376165 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376176 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376188 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376202 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376213 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376224 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376235 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376247 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376260 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376270 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376282 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376300 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376311 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376322 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376334 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376345 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376356 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376368 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376380 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376392 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376403 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376413 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376424 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.376436 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380684 4815 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380720 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380736 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380752 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380766 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380778 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380790 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380803 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380816 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380828 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380842 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380855 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380868 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380880 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380897 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380910 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380924 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380936 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380950 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380964 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380976 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.380988 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381001 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381013 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381025 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381037 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381048 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381061 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381074 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381086 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381099 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381110 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381122 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381134 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381145 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381159 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381171 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381183 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381196 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381206 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381218 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381229 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381241 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381253 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381263 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381274 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381286 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381298 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381309 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381319 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381330 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381342 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381354 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381368 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381381 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381392 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381406 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381419 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381431 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381445 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381457 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381471 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381483 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381514 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381529 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381540 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381552 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381566 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381578 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381591 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381603 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381614 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381628 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381639 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381653 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381665 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381676 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381687 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381699 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381712 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381723 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381735 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381747 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381760 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381772 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381782 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381793 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381805 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381821 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381833 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381844 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381854 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381865 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381876 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381887 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381897 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381909 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381920 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381931 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381944 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381956 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381968 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381980 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.381990 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382001 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382013 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382025 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382036 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382047 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382057 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382068 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382079 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382090 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382101 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382144 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382158 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382169 4815 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382180 4815 reconstruct.go:97] "Volume reconstruction finished" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.382188 4815 reconciler.go:26] "Reconciler: start to sync state" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.394217 4815 manager.go:324] Recovery completed Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.406865 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.409831 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.409877 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.409886 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.410776 4815 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.410804 4815 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.410832 4815 state_mem.go:36] "Initialized new in-memory state store" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.414873 4815 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.417274 4815 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.417340 4815 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.417399 4815 kubelet.go:2335] "Starting kubelet main sync loop" Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.417486 4815 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 05 09:06:41 crc kubenswrapper[4815]: W1205 09:06:41.452058 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.67:6443: connect: connection refused Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.452164 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.67:6443: connect: connection refused" logger="UnhandledError" Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.456878 4815 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.457936 4815 policy_none.go:49] "None policy: Start" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.460412 4815 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.460448 4815 state_mem.go:35] "Initializing new in-memory state store" Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.518544 4815 kubelet.go:2359] "Skipping pod synchronization" err="container runtime status check may not have completed yet" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.524604 4815 manager.go:334] "Starting Device Plugin manager" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.524650 4815 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.524661 4815 server.go:79] "Starting device plugin registration server" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.525061 4815 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.525076 4815 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.525218 4815 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.525356 4815 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.525369 4815 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.535356 4815 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.557015 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" interval="400ms" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.625616 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.627015 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.627045 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.627053 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.627072 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.627470 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.67:6443: connect: connection refused" node="crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.719415 4815 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.719588 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.720783 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.720810 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.720822 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.720937 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.721731 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.721773 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.722515 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.722705 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.722736 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.722751 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.723275 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.723476 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.723998 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.724109 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.724996 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.725003 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.725040 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.725053 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.726223 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.726282 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.726301 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.726585 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.726783 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.726852 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.727712 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.727763 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.727781 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.727921 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.728085 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.728144 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.728089 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.728253 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.728270 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.729032 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.729067 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.729079 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.729276 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.729319 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.729916 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.729949 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.729959 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.730142 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.730166 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.730179 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.786790 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.786865 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.786918 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.786955 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.786984 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.787013 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.787070 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.787322 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.787453 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.787735 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.787842 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.788006 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.788114 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.788222 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.788633 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.827679 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.830137 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.830183 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.830220 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.830248 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.830796 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.67:6443: connect: connection refused" node="crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.890743 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.890792 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.890832 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.890850 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.890864 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.890928 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.890954 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.890982 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891011 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891019 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891071 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891040 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891112 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891125 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891141 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891156 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891254 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891213 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891282 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891299 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891307 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891313 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891360 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891334 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891397 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891204 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891366 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891233 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891468 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: I1205 09:06:41.891590 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 05 09:06:41 crc kubenswrapper[4815]: E1205 09:06:41.958815 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" interval="800ms" Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.059363 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.074966 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 05 09:06:42 crc kubenswrapper[4815]: W1205 09:06:42.092758 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-60e3c08d75c1faf3d5a63f47793bdb8313a2b88925eb000cdac361b4df8130b1 WatchSource:0}: Error finding container 60e3c08d75c1faf3d5a63f47793bdb8313a2b88925eb000cdac361b4df8130b1: Status 404 returned error can't find the container with id 60e3c08d75c1faf3d5a63f47793bdb8313a2b88925eb000cdac361b4df8130b1 Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.098052 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:42 crc kubenswrapper[4815]: W1205 09:06:42.102462 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-3ddb687652a7f85e3bef92951f0877896f858b3df7108690ea9f421a7b80fca0 WatchSource:0}: Error finding container 3ddb687652a7f85e3bef92951f0877896f858b3df7108690ea9f421a7b80fca0: Status 404 returned error can't find the container with id 3ddb687652a7f85e3bef92951f0877896f858b3df7108690ea9f421a7b80fca0 Dec 05 09:06:42 crc kubenswrapper[4815]: W1205 09:06:42.120340 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-d0b2d2aaa4e9376c26e932bb36f572cc4b8b255b6b2998a86c7c8e0a1573c53d WatchSource:0}: Error finding container d0b2d2aaa4e9376c26e932bb36f572cc4b8b255b6b2998a86c7c8e0a1573c53d: Status 404 returned error can't find the container with id d0b2d2aaa4e9376c26e932bb36f572cc4b8b255b6b2998a86c7c8e0a1573c53d Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.125098 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.134670 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 09:06:42 crc kubenswrapper[4815]: W1205 09:06:42.152213 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-94ad32db01e18db804ab691b542c3789822ff827cdb8889715bc78ec2b21aec9 WatchSource:0}: Error finding container 94ad32db01e18db804ab691b542c3789822ff827cdb8889715bc78ec2b21aec9: Status 404 returned error can't find the container with id 94ad32db01e18db804ab691b542c3789822ff827cdb8889715bc78ec2b21aec9 Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.231355 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.232975 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.233011 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.233019 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.233060 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 09:06:42 crc kubenswrapper[4815]: E1205 09:06:42.233448 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.67:6443: connect: connection refused" node="crc" Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.350415 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.67:6443: connect: connection refused Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.425820 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d0b2d2aaa4e9376c26e932bb36f572cc4b8b255b6b2998a86c7c8e0a1573c53d"} Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.430285 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3ddb687652a7f85e3bef92951f0877896f858b3df7108690ea9f421a7b80fca0"} Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.431405 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"60e3c08d75c1faf3d5a63f47793bdb8313a2b88925eb000cdac361b4df8130b1"} Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.432281 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"94ad32db01e18db804ab691b542c3789822ff827cdb8889715bc78ec2b21aec9"} Dec 05 09:06:42 crc kubenswrapper[4815]: I1205 09:06:42.433023 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bebd6ef1643a4c9a0824a4c68001ca7b521cc0e0846ca69a2257524304cc43ac"} Dec 05 09:06:42 crc kubenswrapper[4815]: W1205 09:06:42.737049 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.129.56.67:6443: connect: connection refused Dec 05 09:06:42 crc kubenswrapper[4815]: E1205 09:06:42.737167 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.129.56.67:6443: connect: connection refused" logger="UnhandledError" Dec 05 09:06:42 crc kubenswrapper[4815]: W1205 09:06:42.757394 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.129.56.67:6443: connect: connection refused Dec 05 09:06:42 crc kubenswrapper[4815]: E1205 09:06:42.757517 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.129.56.67:6443: connect: connection refused" logger="UnhandledError" Dec 05 09:06:42 crc kubenswrapper[4815]: E1205 09:06:42.760230 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" interval="1.6s" Dec 05 09:06:42 crc kubenswrapper[4815]: W1205 09:06:42.924137 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.129.56.67:6443: connect: connection refused Dec 05 09:06:42 crc kubenswrapper[4815]: E1205 09:06:42.924218 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.129.56.67:6443: connect: connection refused" logger="UnhandledError" Dec 05 09:06:42 crc kubenswrapper[4815]: W1205 09:06:42.942138 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.129.56.67:6443: connect: connection refused Dec 05 09:06:42 crc kubenswrapper[4815]: E1205 09:06:42.942242 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.129.56.67:6443: connect: connection refused" logger="UnhandledError" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.034028 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.035061 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.035087 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.035095 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.035116 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 09:06:43 crc kubenswrapper[4815]: E1205 09:06:43.035447 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.129.56.67:6443: connect: connection refused" node="crc" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.350465 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.129.56.67:6443: connect: connection refused Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.413562 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 05 09:06:43 crc kubenswrapper[4815]: E1205 09:06:43.414934 4815 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.129.56.67:6443: connect: connection refused" logger="UnhandledError" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.437070 4815 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946" exitCode=0 Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.437131 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946"} Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.437231 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.439455 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.439544 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.439570 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.440643 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"815698fa132d64a49887e7988f69e7fbf24bce894d4565d8408f5a8dec09c786"} Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.440695 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2d43960c9656d05594f1ecd47cd049e0ff80eea4abedc716f0c57aab93e031bb"} Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.440709 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8095ac92a0340ac182451a96eb003f748176d6ca94de46d02ad705caee5d7948"} Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.440720 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3a037ee90341ab082b8523235fd47661c15400baae0eaaa3f54340b6f5885962"} Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.440780 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.441999 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.442051 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.442070 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.444532 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86" exitCode=0 Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.444737 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86"} Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.445096 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.451659 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.451726 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.451743 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.453473 4815 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39" exitCode=0 Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.453673 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.453852 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39"} Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.455009 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.455041 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.455058 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.455374 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.456973 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.457008 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.457021 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.462141 4815 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc" exitCode=0 Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.462194 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc"} Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.462315 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.465989 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.466128 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:43 crc kubenswrapper[4815]: I1205 09:06:43.466240 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.465271 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"3392a2d74d29f75e9794efe23bcb2808d70e32561692fbc759a06eb538bf7ad2"} Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.465344 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.466509 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.466539 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.466548 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.468325 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404"} Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.468385 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4"} Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.468402 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00"} Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.468606 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.469556 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.469595 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.469608 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.474154 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c"} Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.474214 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.474217 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af"} Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.474294 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875"} Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.474313 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8"} Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.474322 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece"} Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.474971 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.474998 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.475007 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.475386 4815 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c" exitCode=0 Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.475452 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c"} Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.475462 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.475612 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.476170 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.476197 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.476211 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.476353 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.476379 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.476391 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.562767 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.635868 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.636847 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.636877 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.636887 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:44 crc kubenswrapper[4815]: I1205 09:06:44.636910 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.481721 4815 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7" exitCode=0 Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.481871 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7"} Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.481929 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.481977 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.481976 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.481929 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.482047 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.482874 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.483341 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.483370 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.483384 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.483515 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.483561 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.483582 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.484359 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.484382 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.484394 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.485079 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.485100 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:45 crc kubenswrapper[4815]: I1205 09:06:45.485110 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.491139 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.491192 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.491745 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993"} Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.491827 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1"} Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.491850 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755"} Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.491869 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd"} Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.492233 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.492279 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.492296 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.908793 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.909156 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.910709 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.910763 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.910780 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:46 crc kubenswrapper[4815]: I1205 09:06:46.917637 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.500720 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79"} Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.500781 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.500917 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.502258 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.502327 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.502353 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.502264 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.502662 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.502678 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.573159 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.806068 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Dec 05 09:06:47 crc kubenswrapper[4815]: I1205 09:06:47.882197 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:48 crc kubenswrapper[4815]: I1205 09:06:48.503633 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 09:06:48 crc kubenswrapper[4815]: I1205 09:06:48.504107 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:48 crc kubenswrapper[4815]: I1205 09:06:48.503740 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:48 crc kubenswrapper[4815]: I1205 09:06:48.505327 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:48 crc kubenswrapper[4815]: I1205 09:06:48.505358 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:48 crc kubenswrapper[4815]: I1205 09:06:48.505368 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:48 crc kubenswrapper[4815]: I1205 09:06:48.505637 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:48 crc kubenswrapper[4815]: I1205 09:06:48.505691 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:48 crc kubenswrapper[4815]: I1205 09:06:48.505711 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:48 crc kubenswrapper[4815]: I1205 09:06:48.995197 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.289340 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.289911 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.290150 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.292328 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.292363 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.292374 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.348447 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.506012 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.506116 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.506901 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.506947 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.506964 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.507332 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.507355 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:49 crc kubenswrapper[4815]: I1205 09:06:49.507365 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:50 crc kubenswrapper[4815]: I1205 09:06:50.508855 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:50 crc kubenswrapper[4815]: I1205 09:06:50.510246 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:50 crc kubenswrapper[4815]: I1205 09:06:50.510293 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:50 crc kubenswrapper[4815]: I1205 09:06:50.510307 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:50 crc kubenswrapper[4815]: I1205 09:06:50.519792 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 09:06:50 crc kubenswrapper[4815]: I1205 09:06:50.520094 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:50 crc kubenswrapper[4815]: I1205 09:06:50.521713 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:50 crc kubenswrapper[4815]: I1205 09:06:50.521753 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:50 crc kubenswrapper[4815]: I1205 09:06:50.521763 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:50 crc kubenswrapper[4815]: I1205 09:06:50.882653 4815 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 09:06:50 crc kubenswrapper[4815]: I1205 09:06:50.882781 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:06:51 crc kubenswrapper[4815]: E1205 09:06:51.535457 4815 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 05 09:06:52 crc kubenswrapper[4815]: I1205 09:06:52.146832 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 05 09:06:52 crc kubenswrapper[4815]: I1205 09:06:52.147166 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:52 crc kubenswrapper[4815]: I1205 09:06:52.149195 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:52 crc kubenswrapper[4815]: I1205 09:06:52.149288 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:52 crc kubenswrapper[4815]: I1205 09:06:52.149310 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:52 crc kubenswrapper[4815]: I1205 09:06:52.399180 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 05 09:06:52 crc kubenswrapper[4815]: I1205 09:06:52.514823 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:52 crc kubenswrapper[4815]: I1205 09:06:52.516200 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:52 crc kubenswrapper[4815]: I1205 09:06:52.516257 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:52 crc kubenswrapper[4815]: I1205 09:06:52.516270 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:54 crc kubenswrapper[4815]: I1205 09:06:54.350791 4815 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 05 09:06:54 crc kubenswrapper[4815]: E1205 09:06:54.361261 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 05 09:06:54 crc kubenswrapper[4815]: E1205 09:06:54.638471 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Dec 05 09:06:54 crc kubenswrapper[4815]: W1205 09:06:54.876407 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 05 09:06:54 crc kubenswrapper[4815]: I1205 09:06:54.876522 4815 trace.go:236] Trace[246432436]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 09:06:44.874) (total time: 10001ms): Dec 05 09:06:54 crc kubenswrapper[4815]: Trace[246432436]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:06:54.876) Dec 05 09:06:54 crc kubenswrapper[4815]: Trace[246432436]: [10.001680221s] [10.001680221s] END Dec 05 09:06:54 crc kubenswrapper[4815]: E1205 09:06:54.876550 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 05 09:06:54 crc kubenswrapper[4815]: W1205 09:06:54.936242 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 05 09:06:54 crc kubenswrapper[4815]: I1205 09:06:54.936402 4815 trace.go:236] Trace[1662527489]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 09:06:44.934) (total time: 10001ms): Dec 05 09:06:54 crc kubenswrapper[4815]: Trace[1662527489]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:06:54.936) Dec 05 09:06:54 crc kubenswrapper[4815]: Trace[1662527489]: [10.001527733s] [10.001527733s] END Dec 05 09:06:54 crc kubenswrapper[4815]: E1205 09:06:54.936447 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 05 09:06:55 crc kubenswrapper[4815]: W1205 09:06:55.028525 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 05 09:06:55 crc kubenswrapper[4815]: I1205 09:06:55.028626 4815 trace.go:236] Trace[499502045]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 09:06:45.027) (total time: 10001ms): Dec 05 09:06:55 crc kubenswrapper[4815]: Trace[499502045]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:06:55.028) Dec 05 09:06:55 crc kubenswrapper[4815]: Trace[499502045]: [10.00140191s] [10.00140191s] END Dec 05 09:06:55 crc kubenswrapper[4815]: E1205 09:06:55.028660 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 05 09:06:55 crc kubenswrapper[4815]: W1205 09:06:55.151485 4815 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout Dec 05 09:06:55 crc kubenswrapper[4815]: I1205 09:06:55.151609 4815 trace.go:236] Trace[1780117052]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (05-Dec-2025 09:06:45.149) (total time: 10002ms): Dec 05 09:06:55 crc kubenswrapper[4815]: Trace[1780117052]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (09:06:55.151) Dec 05 09:06:55 crc kubenswrapper[4815]: Trace[1780117052]: [10.002128071s] [10.002128071s] END Dec 05 09:06:55 crc kubenswrapper[4815]: E1205 09:06:55.151632 4815 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Dec 05 09:06:55 crc kubenswrapper[4815]: I1205 09:06:55.390130 4815 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 05 09:06:55 crc kubenswrapper[4815]: I1205 09:06:55.390221 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 05 09:06:55 crc kubenswrapper[4815]: I1205 09:06:55.397285 4815 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 05 09:06:55 crc kubenswrapper[4815]: I1205 09:06:55.397369 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 05 09:06:57 crc kubenswrapper[4815]: I1205 09:06:57.839335 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:57 crc kubenswrapper[4815]: I1205 09:06:57.840442 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:57 crc kubenswrapper[4815]: I1205 09:06:57.840503 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:57 crc kubenswrapper[4815]: I1205 09:06:57.840517 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:57 crc kubenswrapper[4815]: I1205 09:06:57.840545 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 09:06:57 crc kubenswrapper[4815]: E1205 09:06:57.843867 4815 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.000881 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.001478 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.002551 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.002586 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.002595 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.294541 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.294717 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.295661 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.295681 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.295689 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.300433 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.457037 4815 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.533698 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.534918 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.534950 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:06:59 crc kubenswrapper[4815]: I1205 09:06:59.534961 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.040614 4815 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.395259 4815 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.410821 4815 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.412276 4815 csr.go:261] certificate signing request csr-w8rm7 is approved, waiting to be issued Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.427111 4815 csr.go:257] certificate signing request csr-w8rm7 is issued Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.458200 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.458368 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.459323 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.459362 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.459373 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.464408 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.535366 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.536427 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.536466 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.536503 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.686154 4815 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.854810 4815 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.917754 4815 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37664->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.917828 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37664->192.168.126.11:17697: read: connection reset by peer" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.918222 4815 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37668->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.918275 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:37668->192.168.126.11:17697: read: connection reset by peer" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.919538 4815 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.919590 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.919888 4815 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 05 09:07:00 crc kubenswrapper[4815]: I1205 09:07:00.919923 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.271339 4815 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 05 09:07:01 crc kubenswrapper[4815]: W1205 09:07:01.271604 4815 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Dec 05 09:07:01 crc kubenswrapper[4815]: W1205 09:07:01.271923 4815 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Node ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.355408 4815 apiserver.go:52] "Watching apiserver" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.357557 4815 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.357806 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-ltlpb"] Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.358145 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.358226 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.358253 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.358267 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.358421 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.358514 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.358564 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.358587 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.358725 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ltlpb" Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.358755 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.361541 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.361672 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.362294 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.363746 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.363765 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.363838 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.363974 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.364196 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.364311 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.364383 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.364544 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.364569 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.374894 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.384441 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.393342 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.409753 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.426238 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.428328 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-12-05 09:02:00 +0000 UTC, rotation deadline is 2026-09-16 23:55:23.234823651 +0000 UTC Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.428409 4815 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6854h48m21.806418423s for next certificate rotation Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.436312 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.446300 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.454774 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.455547 4815 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.462734 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.470538 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.476700 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.485028 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.492743 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.500789 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.512508 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.517855 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.517906 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.517929 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.517950 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.517968 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.517987 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518006 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518027 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518051 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518071 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518096 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518117 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518137 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518161 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518183 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518203 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518223 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518245 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518267 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518290 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518314 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518302 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518336 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518358 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518378 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518421 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518441 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518460 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518511 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518521 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518533 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518554 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518574 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518598 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518619 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518638 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518659 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518662 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518691 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.518712 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519029 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519050 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519088 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519110 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519130 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519154 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519177 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519201 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519438 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519611 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519619 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519638 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519664 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519685 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519707 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519728 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519749 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519771 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519791 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519793 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519824 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519846 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519866 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519886 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519908 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519930 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519932 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.519995 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520018 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520041 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520062 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520082 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520101 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520122 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520141 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520159 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520177 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520196 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520214 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520233 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520259 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520279 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520300 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520320 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520344 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520362 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520382 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520399 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520416 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520435 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520453 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520470 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520458 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520677 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520817 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520849 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.520843 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.521006 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.521137 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.521178 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.521436 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.521630 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.521794 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.522572 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.522596 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.522799 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.522881 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.522908 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.522798 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.523138 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.523029 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.523193 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.523278 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.523371 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.523509 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.523716 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.523760 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.523647 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.523821 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.523869 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524045 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524100 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524272 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524383 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524398 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524547 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524605 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524787 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524751 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524889 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524939 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.524987 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525107 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525124 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525165 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525181 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525281 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.525302 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:07:02.025277649 +0000 UTC m=+20.903884486 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525419 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525524 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525562 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525579 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525601 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525620 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525641 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.522890 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525662 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525681 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525700 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525720 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525741 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525759 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525778 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525795 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525813 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525833 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525876 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525895 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525914 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525934 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525953 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525971 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525990 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526009 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526027 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526047 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526065 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526083 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526106 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526123 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526142 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526160 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526199 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526224 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526243 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526262 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526281 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526300 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526320 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526337 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526355 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526374 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526391 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526413 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526440 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526458 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526477 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.526516 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.529457 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.529781 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.525303 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530571 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530607 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530618 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530631 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530652 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530689 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530704 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530715 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530711 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530796 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530823 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530856 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530867 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530883 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530896 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530902 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530943 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530964 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.530983 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531002 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531035 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531096 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531057 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531158 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531178 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531198 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531218 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531239 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531261 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531312 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531322 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531394 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531418 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531439 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.532870 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.532902 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.532922 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.532942 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.532971 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.532990 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.533008 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.533029 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.533049 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.533070 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.533090 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.533131 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.533164 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.533279 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535276 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535340 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535361 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535385 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535431 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535450 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535504 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535526 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535543 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535592 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535609 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535630 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535670 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535688 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535710 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535745 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535764 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535782 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535815 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535840 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535870 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535913 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535935 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535959 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535977 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536002 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/818e2dfa-a1d7-403d-a475-5437ece116cb-hosts-file\") pod \"node-resolver-ltlpb\" (UID: \"818e2dfa-a1d7-403d-a475-5437ece116cb\") " pod="openshift-dns/node-resolver-ltlpb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536021 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536040 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536070 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvv88\" (UniqueName: \"kubernetes.io/projected/818e2dfa-a1d7-403d-a475-5437ece116cb-kube-api-access-vvv88\") pod \"node-resolver-ltlpb\" (UID: \"818e2dfa-a1d7-403d-a475-5437ece116cb\") " pod="openshift-dns/node-resolver-ltlpb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536092 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536119 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536138 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536153 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536171 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536191 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536395 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536412 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536483 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536520 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536532 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536543 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536556 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536567 4815 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536579 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536618 4815 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536634 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536644 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536655 4815 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536665 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536699 4815 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536709 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536718 4815 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536727 4815 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536737 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536747 4815 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536778 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536788 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536797 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536808 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536822 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536849 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536861 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536870 4815 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536880 4815 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536889 4815 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536898 4815 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536907 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536943 4815 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536957 4815 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536970 4815 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.536983 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537017 4815 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537032 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537047 4815 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537056 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537066 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537075 4815 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537083 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537092 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537101 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537110 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537120 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537130 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537139 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537149 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537162 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537171 4815 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537180 4815 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537189 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537198 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537208 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537217 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537226 4815 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537237 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537249 4815 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537268 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537283 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537296 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537309 4815 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.542108 4815 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.554596 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.557391 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.564864 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.567046 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.567478 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531551 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531664 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.531841 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.532258 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.532307 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.532637 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.532824 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.532931 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.533195 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.533556 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.534070 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.534112 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.534643 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535160 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535166 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535222 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.535381 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537270 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537538 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537680 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537836 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.537970 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.538610 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.538751 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.538823 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.539281 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.539842 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.540151 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.540184 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.543094 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.543237 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.543347 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.543755 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.544029 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.544734 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.544913 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.545385 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.545602 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.545854 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.549158 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.549283 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.549655 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.550262 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.551288 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.551782 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.552675 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.552682 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.554250 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.554266 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.554327 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.554846 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.555095 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.555630 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.557293 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.557486 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.557741 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.560013 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.560355 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.560628 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.561255 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.562133 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.562412 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.562656 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.562761 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.562770 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.562959 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.563082 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.563138 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.563313 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.563533 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.563547 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.563574 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.563744 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.563941 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.564187 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.564311 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.564630 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.565393 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.566138 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.566919 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.567658 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.568583 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.569667 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.569763 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.571333 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.571909 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.572057 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.572169 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.572220 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.572330 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.572428 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.572640 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.574668 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.575202 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.575225 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.590025 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.590391 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.575330 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.575452 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.575477 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.575548 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.575719 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.575741 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.590855 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:02.090818714 +0000 UTC m=+20.969425541 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.575804 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.575961 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.576051 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.576287 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.576642 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.576476 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.579329 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.579672 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.579952 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.586769 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.587042 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.587071 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.587298 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.587307 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.587517 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.587592 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.587627 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.588136 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.588176 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.588191 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.591100 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:02.091088382 +0000 UTC m=+20.969695219 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.591440 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.591925 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.592269 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.592632 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.592983 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.593275 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.593563 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.593882 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.595871 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.597828 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.614224 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.614668 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.614827 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.615071 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:02.11504815 +0000 UTC m=+20.993654987 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.617260 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c" exitCode=255 Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.617324 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c"} Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.619624 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.619666 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.619679 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:01 crc kubenswrapper[4815]: E1205 09:07:01.619741 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:02.119711563 +0000 UTC m=+20.998318400 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.624449 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.635873 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.636366 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640386 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640434 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/818e2dfa-a1d7-403d-a475-5437ece116cb-hosts-file\") pod \"node-resolver-ltlpb\" (UID: \"818e2dfa-a1d7-403d-a475-5437ece116cb\") " pod="openshift-dns/node-resolver-ltlpb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640457 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640474 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvv88\" (UniqueName: \"kubernetes.io/projected/818e2dfa-a1d7-403d-a475-5437ece116cb-kube-api-access-vvv88\") pod \"node-resolver-ltlpb\" (UID: \"818e2dfa-a1d7-403d-a475-5437ece116cb\") " pod="openshift-dns/node-resolver-ltlpb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640539 4815 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640550 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640559 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640567 4815 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640576 4815 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640584 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640594 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640601 4815 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640611 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640619 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640627 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640641 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640650 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640660 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640670 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640683 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640692 4815 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640701 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640709 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640716 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640725 4815 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640733 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640741 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640749 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640756 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640765 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640773 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640780 4815 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640790 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640800 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640810 4815 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640819 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640828 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640837 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640847 4815 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640855 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640863 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640872 4815 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640882 4815 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640891 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640900 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640910 4815 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640920 4815 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640930 4815 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640939 4815 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640949 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640959 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640968 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641029 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.640977 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641104 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641085 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641117 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/818e2dfa-a1d7-403d-a475-5437ece116cb-hosts-file\") pod \"node-resolver-ltlpb\" (UID: \"818e2dfa-a1d7-403d-a475-5437ece116cb\") " pod="openshift-dns/node-resolver-ltlpb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641114 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641191 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641210 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641256 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641268 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641279 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641290 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641301 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641310 4815 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641320 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641330 4815 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641340 4815 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641350 4815 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641359 4815 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641368 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641378 4815 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641388 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641398 4815 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641410 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641419 4815 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641428 4815 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641438 4815 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641447 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641459 4815 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641469 4815 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641497 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641508 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641518 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641528 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641537 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641546 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641556 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641566 4815 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641576 4815 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641585 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641595 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641605 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641615 4815 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641626 4815 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641636 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641646 4815 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641656 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641668 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641679 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641689 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641700 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641711 4815 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641721 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641731 4815 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641740 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641750 4815 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641760 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641770 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641780 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641804 4815 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641821 4815 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641832 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641841 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641853 4815 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641863 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641880 4815 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641890 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641901 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641910 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641920 4815 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641930 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641941 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641951 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641962 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641972 4815 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641983 4815 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.641994 4815 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.642005 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.642016 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.642027 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.642036 4815 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.642046 4815 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.642055 4815 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.642065 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.642074 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.642083 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.642092 4815 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.646756 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.657073 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.659941 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvv88\" (UniqueName: \"kubernetes.io/projected/818e2dfa-a1d7-403d-a475-5437ece116cb-kube-api-access-vvv88\") pod \"node-resolver-ltlpb\" (UID: \"818e2dfa-a1d7-403d-a475-5437ece116cb\") " pod="openshift-dns/node-resolver-ltlpb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.667076 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.674205 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.674668 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.682195 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-ltlpb" Dec 05 09:07:01 crc kubenswrapper[4815]: W1205 09:07:01.686123 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-7517559353bc1a28442917cd485224eda3946a3d579b92ced59e16c13c908056 WatchSource:0}: Error finding container 7517559353bc1a28442917cd485224eda3946a3d579b92ced59e16c13c908056: Status 404 returned error can't find the container with id 7517559353bc1a28442917cd485224eda3946a3d579b92ced59e16c13c908056 Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.686204 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.691082 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.696940 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.697085 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.707370 4815 scope.go:117] "RemoveContainer" containerID="faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c" Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.707914 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 09:07:01 crc kubenswrapper[4815]: I1205 09:07:01.708187 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.045732 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.045873 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:07:03.04585828 +0000 UTC m=+21.924465117 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.146633 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.146685 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.146705 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.146727 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.146903 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.146925 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.146937 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.146975 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.147007 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.147005 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.147018 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.146983 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:03.146969441 +0000 UTC m=+22.025576278 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.147164 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:03.147147336 +0000 UTC m=+22.025754173 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.147175 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:03.147169747 +0000 UTC m=+22.025776584 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.147283 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:02 crc kubenswrapper[4815]: E1205 09:07:02.147328 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:03.147321171 +0000 UTC m=+22.025928008 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.156373 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-vrs47"] Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.156907 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.247947 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-system-cni-dir\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.247997 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-os-release\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.248056 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9a27419f-8364-4ff9-a6bd-677c3fa72fef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.248206 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.248301 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clbvz\" (UniqueName: \"kubernetes.io/projected/9a27419f-8364-4ff9-a6bd-677c3fa72fef-kube-api-access-clbvz\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.248400 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-cnibin\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.248551 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9a27419f-8364-4ff9-a6bd-677c3fa72fef-cni-binary-copy\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.255326 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.255759 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.256953 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.259865 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.259922 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.261321 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-4wq7d"] Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.261887 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.269038 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.269542 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.270395 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-f9lkk"] Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.272792 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.273345 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.273647 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.275038 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.276825 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.277853 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.288891 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.301741 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.310330 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.318122 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.331477 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.346475 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.349406 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-run-k8s-cni-cncf-io\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.349450 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dff16910-bb36-4748-82bb-0f1d90da05b5-cni-binary-copy\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.349476 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv78q\" (UniqueName: \"kubernetes.io/projected/dff16910-bb36-4748-82bb-0f1d90da05b5-kube-api-access-tv78q\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.349535 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-conf-dir\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.349558 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-os-release\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.349578 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-run-netns\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.349646 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-var-lib-cni-multus\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350298 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9a27419f-8364-4ff9-a6bd-677c3fa72fef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350365 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-cnibin\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350385 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-run-multus-certs\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350405 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0a874c44-831d-4f81-bf48-2025f5615574-mcd-auth-proxy-config\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350420 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-system-cni-dir\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350436 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350453 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-cni-dir\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350474 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-var-lib-kubelet\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350536 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-etc-kubernetes\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350558 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0a874c44-831d-4f81-bf48-2025f5615574-proxy-tls\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350572 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-var-lib-cni-bin\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350597 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-cnibin\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350613 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9a27419f-8364-4ff9-a6bd-677c3fa72fef-cni-binary-copy\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350669 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-os-release\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350709 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0a874c44-831d-4f81-bf48-2025f5615574-rootfs\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350728 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxzq4\" (UniqueName: \"kubernetes.io/projected/0a874c44-831d-4f81-bf48-2025f5615574-kube-api-access-fxzq4\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350759 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-system-cni-dir\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350776 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-socket-dir-parent\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350793 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-hostroot\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350806 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-daemon-config\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.350825 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clbvz\" (UniqueName: \"kubernetes.io/projected/9a27419f-8364-4ff9-a6bd-677c3fa72fef-kube-api-access-clbvz\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.351251 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/9a27419f-8364-4ff9-a6bd-677c3fa72fef-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.351285 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-cnibin\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.351408 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-os-release\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.351474 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-system-cni-dir\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.351808 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/9a27419f-8364-4ff9-a6bd-677c3fa72fef-cni-binary-copy\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.360082 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/9a27419f-8364-4ff9-a6bd-677c3fa72fef-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.361212 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.369697 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clbvz\" (UniqueName: \"kubernetes.io/projected/9a27419f-8364-4ff9-a6bd-677c3fa72fef-kube-api-access-clbvz\") pod \"multus-additional-cni-plugins-vrs47\" (UID: \"9a27419f-8364-4ff9-a6bd-677c3fa72fef\") " pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.373348 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.394209 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.406185 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.417932 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.432592 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.434580 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.449699 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.450281 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451163 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-cni-dir\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451218 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-var-lib-kubelet\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451263 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-etc-kubernetes\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451298 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0a874c44-831d-4f81-bf48-2025f5615574-proxy-tls\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451307 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-cni-dir\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451328 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-var-lib-cni-bin\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451379 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-etc-kubernetes\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451406 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0a874c44-831d-4f81-bf48-2025f5615574-rootfs\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451431 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-var-lib-cni-bin\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451436 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxzq4\" (UniqueName: \"kubernetes.io/projected/0a874c44-831d-4f81-bf48-2025f5615574-kube-api-access-fxzq4\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451457 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/0a874c44-831d-4f81-bf48-2025f5615574-rootfs\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451468 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-socket-dir-parent\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451532 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-hostroot\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451563 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-daemon-config\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451592 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-run-k8s-cni-cncf-io\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451635 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dff16910-bb36-4748-82bb-0f1d90da05b5-cni-binary-copy\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451668 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv78q\" (UniqueName: \"kubernetes.io/projected/dff16910-bb36-4748-82bb-0f1d90da05b5-kube-api-access-tv78q\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451701 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-os-release\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451729 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-conf-dir\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451761 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-run-netns\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451789 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-var-lib-cni-multus\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451818 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0a874c44-831d-4f81-bf48-2025f5615574-mcd-auth-proxy-config\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451846 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-system-cni-dir\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451868 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-os-release\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451873 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-cnibin\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451916 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-run-multus-certs\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451941 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-cnibin\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451953 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-run-multus-certs\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451634 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-socket-dir-parent\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.451999 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-conf-dir\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.452052 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-run-netns\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.452096 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-var-lib-cni-multus\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.452174 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-run-k8s-cni-cncf-io\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.452201 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-hostroot\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.452234 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-system-cni-dir\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.452346 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/dff16910-bb36-4748-82bb-0f1d90da05b5-cni-binary-copy\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.452371 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/dff16910-bb36-4748-82bb-0f1d90da05b5-multus-daemon-config\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.452387 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/dff16910-bb36-4748-82bb-0f1d90da05b5-host-var-lib-kubelet\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.453178 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0a874c44-831d-4f81-bf48-2025f5615574-mcd-auth-proxy-config\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.453240 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.455947 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0a874c44-831d-4f81-bf48-2025f5615574-proxy-tls\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.468975 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxzq4\" (UniqueName: \"kubernetes.io/projected/0a874c44-831d-4f81-bf48-2025f5615574-kube-api-access-fxzq4\") pod \"machine-config-daemon-4wq7d\" (UID: \"0a874c44-831d-4f81-bf48-2025f5615574\") " pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.470667 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.471695 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vrs47" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.472625 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv78q\" (UniqueName: \"kubernetes.io/projected/dff16910-bb36-4748-82bb-0f1d90da05b5-kube-api-access-tv78q\") pod \"multus-f9lkk\" (UID: \"dff16910-bb36-4748-82bb-0f1d90da05b5\") " pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: W1205 09:07:02.481232 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a27419f_8364_4ff9_a6bd_677c3fa72fef.slice/crio-ddec2ace5c22aa2c14d195a09f99b053fe44f847b0db6911e03afa3792ce5cee WatchSource:0}: Error finding container ddec2ace5c22aa2c14d195a09f99b053fe44f847b0db6911e03afa3792ce5cee: Status 404 returned error can't find the container with id ddec2ace5c22aa2c14d195a09f99b053fe44f847b0db6911e03afa3792ce5cee Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.481687 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.494435 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.504831 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.529541 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.538826 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.549091 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.561555 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.580028 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.587552 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.593522 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.597784 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-f9lkk" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.617392 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.623464 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7517559353bc1a28442917cd485224eda3946a3d579b92ced59e16c13c908056"} Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.627770 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pzlxw"] Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.629005 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.633325 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.633633 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.633977 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.634519 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.634865 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.635196 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.635300 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.643500 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.645777 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a"} Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.645825 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd"} Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.645838 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c586c092236e6e4d92b99dd2ace9e35bee38e999c4009df133b7cbd26fecbeae"} Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.651661 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ltlpb" event={"ID":"818e2dfa-a1d7-403d-a475-5437ece116cb","Type":"ContainerStarted","Data":"925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7"} Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.651700 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-ltlpb" event={"ID":"818e2dfa-a1d7-403d-a475-5437ece116cb","Type":"ContainerStarted","Data":"0d4d518b55b623d5ed1bbccb1cc46e3f497dfa55b200f3ffe50c2bc956ec41a2"} Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.654911 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.656011 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9"} Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.656159 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.656606 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"313fde87754cbdcbb93a6688f961bb4cb93753bb1699e668a540c2b4ed61b07f"} Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.657120 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" event={"ID":"9a27419f-8364-4ff9-a6bd-677c3fa72fef","Type":"ContainerStarted","Data":"ddec2ace5c22aa2c14d195a09f99b053fe44f847b0db6911e03afa3792ce5cee"} Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.658188 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11"} Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.658212 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"c381ed431d4d641a3c3a0c421a3079ee9b15de9801174d25a7a74ccd017dab31"} Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.663469 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.681411 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.695289 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.718685 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.740060 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755044 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-var-lib-openvswitch\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755099 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-kubelet\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755128 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-netns\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755612 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-bin\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755803 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-config\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755833 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e930cfad-5770-4db6-826b-353554e2f23c-ovn-node-metrics-cert\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755866 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-etc-openvswitch\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755886 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-script-lib\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755906 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-openvswitch\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755924 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-env-overrides\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755967 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-systemd\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.755987 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-ovn\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.756019 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-systemd-units\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.756038 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-slash\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.756431 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.756577 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgqrd\" (UniqueName: \"kubernetes.io/projected/e930cfad-5770-4db6-826b-353554e2f23c-kube-api-access-bgqrd\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.756632 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-ovn-kubernetes\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.756675 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-netd\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.756720 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-node-log\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.756787 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-log-socket\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.758016 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.769091 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.780750 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.798427 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.811984 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.822448 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.833647 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.851462 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.857924 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-etc-openvswitch\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.857957 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-script-lib\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.857980 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-openvswitch\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.857994 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-env-overrides\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858008 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-systemd-units\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858022 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-slash\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858036 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-systemd\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858050 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-ovn\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858064 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858080 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgqrd\" (UniqueName: \"kubernetes.io/projected/e930cfad-5770-4db6-826b-353554e2f23c-kube-api-access-bgqrd\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858096 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-ovn-kubernetes\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858104 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-systemd-units\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858116 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-etc-openvswitch\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858151 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-slash\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858129 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-openvswitch\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858112 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-node-log\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858181 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-systemd\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858133 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-node-log\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858199 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-netd\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858222 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-ovn\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858234 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-log-socket\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858253 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-ovn-kubernetes\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858277 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-var-lib-openvswitch\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858294 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-netd\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858301 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-kubelet\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858159 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858335 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-netns\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858362 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-bin\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858377 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-kubelet\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858383 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-netns\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858336 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-log-socket\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858380 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-config\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858408 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-bin\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858419 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e930cfad-5770-4db6-826b-353554e2f23c-ovn-node-metrics-cert\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858363 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-var-lib-openvswitch\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858732 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-env-overrides\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858859 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-script-lib\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.858903 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-config\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.865337 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.878169 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.895120 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.907162 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.920170 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.925445 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgqrd\" (UniqueName: \"kubernetes.io/projected/e930cfad-5770-4db6-826b-353554e2f23c-kube-api-access-bgqrd\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.926068 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e930cfad-5770-4db6-826b-353554e2f23c-ovn-node-metrics-cert\") pod \"ovnkube-node-pzlxw\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.939431 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.948139 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:02 crc kubenswrapper[4815]: W1205 09:07:02.961879 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode930cfad_5770_4db6_826b_353554e2f23c.slice/crio-9ce7716475da15fbe3f700791633122ab8e9d196e7a038a6ab2483390928fe29 WatchSource:0}: Error finding container 9ce7716475da15fbe3f700791633122ab8e9d196e7a038a6ab2483390928fe29: Status 404 returned error can't find the container with id 9ce7716475da15fbe3f700791633122ab8e9d196e7a038a6ab2483390928fe29 Dec 05 09:07:02 crc kubenswrapper[4815]: I1205 09:07:02.966179 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:02Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.061174 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.061375 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:07:05.061358942 +0000 UTC m=+23.939965779 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.161929 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.161973 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.162001 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.162030 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162101 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162145 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:05.162132184 +0000 UTC m=+24.040739021 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162267 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162326 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162351 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162464 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:05.162439263 +0000 UTC m=+24.041046100 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162526 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162542 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162552 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162565 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162579 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:05.162571806 +0000 UTC m=+24.041178633 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.162596 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:05.162588827 +0000 UTC m=+24.041195664 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.418501 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.418614 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.418969 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.419041 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.419147 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:03 crc kubenswrapper[4815]: E1205 09:07:03.419223 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.425755 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.426459 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.427574 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.428175 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.429141 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.429654 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.430217 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.433918 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.434678 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.435613 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.436389 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.437897 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.438411 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.438901 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.439831 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.440336 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.441319 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.441748 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.442266 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.443337 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.443808 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.444923 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.445333 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.446412 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.446909 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.447559 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.448722 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.449157 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.450119 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.450639 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.451505 4815 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.451603 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.453376 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.454412 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.454993 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.457555 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.458622 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.459696 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.460651 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.461941 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.462657 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.464301 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.464980 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.465955 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.466392 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.467259 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.467951 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.469069 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.469642 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.470463 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.470915 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.471830 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.472390 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.473031 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.661467 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a" exitCode=0 Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.661556 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a"} Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.661586 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"9ce7716475da15fbe3f700791633122ab8e9d196e7a038a6ab2483390928fe29"} Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.664579 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-f9lkk" event={"ID":"dff16910-bb36-4748-82bb-0f1d90da05b5","Type":"ContainerStarted","Data":"f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded"} Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.664707 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-f9lkk" event={"ID":"dff16910-bb36-4748-82bb-0f1d90da05b5","Type":"ContainerStarted","Data":"44baef633a87be28d7ef00b7c0afea6a5414065166416da041b2dc69c3dcfa25"} Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.666438 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512"} Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.666530 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2"} Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.668139 4815 generic.go:334] "Generic (PLEG): container finished" podID="9a27419f-8364-4ff9-a6bd-677c3fa72fef" containerID="e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292" exitCode=0 Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.669200 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" event={"ID":"9a27419f-8364-4ff9-a6bd-677c3fa72fef","Type":"ContainerDied","Data":"e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292"} Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.701612 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.751579 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.774032 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.795098 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.853128 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.873877 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.892729 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.905828 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.918687 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.933321 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.949556 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.964431 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.978582 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:03 crc kubenswrapper[4815]: I1205 09:07:03.991712 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:03Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.003390 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.016790 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.031447 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.048550 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.067713 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.086185 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.112560 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.129618 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.147864 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.173284 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.186429 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.197119 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.244131 4815 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.245889 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.245919 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.245928 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.246018 4815 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.252330 4815 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.252616 4815 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.253544 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.253572 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.253584 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.253599 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.253611 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: E1205 09:07:04.277467 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.280500 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.280535 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.280544 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.280560 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.280570 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: E1205 09:07:04.293273 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.296812 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.296834 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.296844 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.296856 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.296865 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: E1205 09:07:04.309238 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.312406 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.312448 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.312461 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.312478 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.312511 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: E1205 09:07:04.323886 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.327903 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.327934 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.327943 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.327957 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.327966 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: E1205 09:07:04.338963 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: E1205 09:07:04.339077 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.340362 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.340393 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.340415 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.340430 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.340440 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.442841 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.442869 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.442876 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.442889 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.442898 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.545881 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.545935 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.545953 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.545970 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.545997 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.649059 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.649117 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.649136 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.649165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.649188 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.674602 4815 generic.go:334] "Generic (PLEG): container finished" podID="9a27419f-8364-4ff9-a6bd-677c3fa72fef" containerID="9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b" exitCode=0 Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.674772 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" event={"ID":"9a27419f-8364-4ff9-a6bd-677c3fa72fef","Type":"ContainerDied","Data":"9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.677192 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.687735 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.687805 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.687824 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.687840 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.687857 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.687872 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.696596 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.712110 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.726132 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.741395 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.751936 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.751972 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.751983 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.751999 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.752012 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.755023 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.765616 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.777561 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.788964 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.803629 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.816679 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.837690 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.853822 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.854028 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.854041 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.854054 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.854062 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.857398 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.871244 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.884968 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.900469 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.913464 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.927632 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.938171 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.951342 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.955672 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.955697 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.955705 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.955717 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.955728 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:04Z","lastTransitionTime":"2025-12-05T09:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.964666 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:04 crc kubenswrapper[4815]: I1205 09:07:04.984747 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:04Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.003239 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.016714 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.035483 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.049596 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.057916 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.057960 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.057971 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.057987 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.057998 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:05Z","lastTransitionTime":"2025-12-05T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.061567 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.080224 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.080412 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:07:09.080391071 +0000 UTC m=+27.958997908 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.160435 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.160517 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.160533 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.160552 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.160567 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:05Z","lastTransitionTime":"2025-12-05T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.181301 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.181350 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.181381 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.181402 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181449 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181523 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:09.181506432 +0000 UTC m=+28.060113269 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181532 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181544 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181548 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181559 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181564 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181570 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181596 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:09.181586774 +0000 UTC m=+28.060193611 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181614 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:09.181606614 +0000 UTC m=+28.060213461 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181896 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.181951 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:09.181937744 +0000 UTC m=+28.060544581 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.263642 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.263683 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.263695 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.263711 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.263722 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:05Z","lastTransitionTime":"2025-12-05T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.365762 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.365808 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.365819 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.365837 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.365852 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:05Z","lastTransitionTime":"2025-12-05T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.418152 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.418256 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.418162 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.418400 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.418271 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:05 crc kubenswrapper[4815]: E1205 09:07:05.418530 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.495693 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.495720 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.495728 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.495740 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.495749 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:05Z","lastTransitionTime":"2025-12-05T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.598403 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.598681 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.598795 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.598884 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.598992 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:05Z","lastTransitionTime":"2025-12-05T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.695685 4815 generic.go:334] "Generic (PLEG): container finished" podID="9a27419f-8364-4ff9-a6bd-677c3fa72fef" containerID="938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350" exitCode=0 Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.697403 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" event={"ID":"9a27419f-8364-4ff9-a6bd-677c3fa72fef","Type":"ContainerDied","Data":"938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350"} Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.703263 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.703301 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.703309 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.703328 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.703338 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:05Z","lastTransitionTime":"2025-12-05T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.722647 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.747875 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.765288 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.779062 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.795555 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.809531 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.809571 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.809579 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.809596 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.809608 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:05Z","lastTransitionTime":"2025-12-05T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.810768 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.824576 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.838547 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.855618 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.878338 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.893872 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.911921 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.912032 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.912042 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.912062 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.912072 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:05Z","lastTransitionTime":"2025-12-05T09:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.916056 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:05 crc kubenswrapper[4815]: I1205 09:07:05.930982 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.014240 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.014279 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.014290 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.014305 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.014315 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:06Z","lastTransitionTime":"2025-12-05T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.117790 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.117832 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.117843 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.117865 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.117879 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:06Z","lastTransitionTime":"2025-12-05T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.219906 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.219945 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.219953 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.219968 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.219977 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:06Z","lastTransitionTime":"2025-12-05T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.323258 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.323995 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.324087 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.324177 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.324252 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:06Z","lastTransitionTime":"2025-12-05T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.427728 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.427766 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.427781 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.427798 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.427811 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:06Z","lastTransitionTime":"2025-12-05T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.447037 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-625kt"] Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.447522 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-625kt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.449845 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.449895 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.451521 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.451667 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.469909 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.489373 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.507369 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.520723 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.530072 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.530109 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.530120 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.530137 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.530148 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:06Z","lastTransitionTime":"2025-12-05T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.537064 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.552801 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.568231 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.582880 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.602227 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.606421 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57-host\") pod \"node-ca-625kt\" (UID: \"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\") " pod="openshift-image-registry/node-ca-625kt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.606518 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljd7h\" (UniqueName: \"kubernetes.io/projected/9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57-kube-api-access-ljd7h\") pod \"node-ca-625kt\" (UID: \"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\") " pod="openshift-image-registry/node-ca-625kt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.606644 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57-serviceca\") pod \"node-ca-625kt\" (UID: \"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\") " pod="openshift-image-registry/node-ca-625kt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.647965 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.649515 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.649546 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.649555 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.649575 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.649586 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:06Z","lastTransitionTime":"2025-12-05T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.662382 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.674901 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.688059 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.707278 4815 generic.go:334] "Generic (PLEG): container finished" podID="9a27419f-8364-4ff9-a6bd-677c3fa72fef" containerID="0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2" exitCode=0 Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.707334 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57-host\") pod \"node-ca-625kt\" (UID: \"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\") " pod="openshift-image-registry/node-ca-625kt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.707369 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" event={"ID":"9a27419f-8364-4ff9-a6bd-677c3fa72fef","Type":"ContainerDied","Data":"0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2"} Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.707392 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljd7h\" (UniqueName: \"kubernetes.io/projected/9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57-kube-api-access-ljd7h\") pod \"node-ca-625kt\" (UID: \"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\") " pod="openshift-image-registry/node-ca-625kt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.707505 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57-serviceca\") pod \"node-ca-625kt\" (UID: \"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\") " pod="openshift-image-registry/node-ca-625kt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.707566 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57-host\") pod \"node-ca-625kt\" (UID: \"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\") " pod="openshift-image-registry/node-ca-625kt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.709066 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57-serviceca\") pod \"node-ca-625kt\" (UID: \"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\") " pod="openshift-image-registry/node-ca-625kt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.713993 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0"} Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.717016 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.740729 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljd7h\" (UniqueName: \"kubernetes.io/projected/9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57-kube-api-access-ljd7h\") pod \"node-ca-625kt\" (UID: \"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\") " pod="openshift-image-registry/node-ca-625kt" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.742239 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.756715 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.756766 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.756779 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.756798 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.756815 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:06Z","lastTransitionTime":"2025-12-05T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.760305 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.762756 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-625kt" Dec 05 09:07:06 crc kubenswrapper[4815]: W1205 09:07:06.776514 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9193a1c8_5c57_4bec_a4cf_3fc1b7e6be57.slice/crio-559f96fa5619e840bedbd61739e523fd9a71eaf83230fa68e3cf8c53162aa0cf WatchSource:0}: Error finding container 559f96fa5619e840bedbd61739e523fd9a71eaf83230fa68e3cf8c53162aa0cf: Status 404 returned error can't find the container with id 559f96fa5619e840bedbd61739e523fd9a71eaf83230fa68e3cf8c53162aa0cf Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.780985 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.797572 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.822001 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.841794 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.858402 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.858430 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.858438 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.858450 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.858460 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:06Z","lastTransitionTime":"2025-12-05T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.861408 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.875053 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.888740 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.901340 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.919737 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.935788 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.948090 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.958218 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.960261 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.960303 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.960312 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.960325 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:06 crc kubenswrapper[4815]: I1205 09:07:06.960334 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:06Z","lastTransitionTime":"2025-12-05T09:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.062906 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.063139 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.063147 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.063163 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.063173 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:07Z","lastTransitionTime":"2025-12-05T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.166048 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.166080 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.166090 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.166104 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.166115 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:07Z","lastTransitionTime":"2025-12-05T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.268241 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.268272 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.268283 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.268297 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.268307 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:07Z","lastTransitionTime":"2025-12-05T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.370798 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.370826 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.370836 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.370850 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.370861 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:07Z","lastTransitionTime":"2025-12-05T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.418720 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:07 crc kubenswrapper[4815]: E1205 09:07:07.418841 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.419216 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:07 crc kubenswrapper[4815]: E1205 09:07:07.419338 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.419459 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:07 crc kubenswrapper[4815]: E1205 09:07:07.419536 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.472981 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.473014 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.473025 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.473041 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.473056 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:07Z","lastTransitionTime":"2025-12-05T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.575547 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.575591 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.575602 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.575620 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.575632 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:07Z","lastTransitionTime":"2025-12-05T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.678686 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.678721 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.678730 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.678745 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.678761 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:07Z","lastTransitionTime":"2025-12-05T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.721472 4815 generic.go:334] "Generic (PLEG): container finished" podID="9a27419f-8364-4ff9-a6bd-677c3fa72fef" containerID="5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df" exitCode=0 Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.721561 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" event={"ID":"9a27419f-8364-4ff9-a6bd-677c3fa72fef","Type":"ContainerDied","Data":"5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.726560 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-625kt" event={"ID":"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57","Type":"ContainerStarted","Data":"24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.726613 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-625kt" event={"ID":"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57","Type":"ContainerStarted","Data":"559f96fa5619e840bedbd61739e523fd9a71eaf83230fa68e3cf8c53162aa0cf"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.746387 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.761511 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.774722 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.781783 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.781847 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.781859 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.781879 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.781893 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:07Z","lastTransitionTime":"2025-12-05T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.787170 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.799900 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.809979 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.820739 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.835475 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.853738 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.880244 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.893168 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.893686 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.893698 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.893719 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.893730 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:07Z","lastTransitionTime":"2025-12-05T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.906283 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.919200 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.933433 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.944802 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.957735 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.969176 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.981252 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.994358 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:07Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.996126 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.996153 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.996162 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.996177 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:07 crc kubenswrapper[4815]: I1205 09:07:07.996187 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:07Z","lastTransitionTime":"2025-12-05T09:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.011413 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.029323 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.044713 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.055561 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.067959 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.078870 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.093474 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.098103 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.098142 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.098151 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.098165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.098175 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:08Z","lastTransitionTime":"2025-12-05T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.109022 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.123152 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.137319 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.200068 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.200109 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.200121 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.200136 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.200148 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:08Z","lastTransitionTime":"2025-12-05T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.301675 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.301724 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.301739 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.301762 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.301778 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:08Z","lastTransitionTime":"2025-12-05T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.403771 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.403814 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.403828 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.403867 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.403878 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:08Z","lastTransitionTime":"2025-12-05T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.506451 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.506531 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.506546 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.506563 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.506574 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:08Z","lastTransitionTime":"2025-12-05T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.609043 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.609071 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.609079 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.609091 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.609100 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:08Z","lastTransitionTime":"2025-12-05T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.712257 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.712308 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.712324 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.712348 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.712364 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:08Z","lastTransitionTime":"2025-12-05T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.733177 4815 generic.go:334] "Generic (PLEG): container finished" podID="9a27419f-8364-4ff9-a6bd-677c3fa72fef" containerID="70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd" exitCode=0 Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.733250 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" event={"ID":"9a27419f-8364-4ff9-a6bd-677c3fa72fef","Type":"ContainerDied","Data":"70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd"} Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.760101 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.781327 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.794299 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.814697 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.815451 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.815473 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.815482 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.815514 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.815524 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:08Z","lastTransitionTime":"2025-12-05T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.823548 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.834925 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.844574 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.857167 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.868122 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.882785 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.893306 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.909674 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.921833 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.921872 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.921882 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.921897 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.921907 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:08Z","lastTransitionTime":"2025-12-05T09:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.924738 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:08 crc kubenswrapper[4815]: I1205 09:07:08.948417 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.023981 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.024014 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.024028 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.024047 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.024059 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:09Z","lastTransitionTime":"2025-12-05T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.127977 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.128015 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.128026 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.128043 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.128054 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:09Z","lastTransitionTime":"2025-12-05T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.130178 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.130358 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:07:17.130340102 +0000 UTC m=+36.008946939 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.230999 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.231081 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.231128 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.231163 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231271 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231348 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231356 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231397 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231411 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231458 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231420 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231517 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231393 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:17.231359449 +0000 UTC m=+36.109966326 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231620 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:17.231592976 +0000 UTC m=+36.110199853 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231659 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:17.231645557 +0000 UTC m=+36.110252424 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.231679 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:17.231668918 +0000 UTC m=+36.110275795 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.232596 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.232636 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.232653 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.232677 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.232694 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:09Z","lastTransitionTime":"2025-12-05T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.322955 4815 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.336008 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.336047 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.336057 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.336072 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.336082 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:09Z","lastTransitionTime":"2025-12-05T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.418351 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.418352 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.418468 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.418357 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.418632 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:09 crc kubenswrapper[4815]: E1205 09:07:09.418706 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.438656 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.438708 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.438723 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.438744 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.438759 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:09Z","lastTransitionTime":"2025-12-05T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.541124 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.541160 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.541171 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.541186 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.541197 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:09Z","lastTransitionTime":"2025-12-05T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.643664 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.643987 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.644006 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.644029 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.644047 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:09Z","lastTransitionTime":"2025-12-05T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.746348 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.746791 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.746941 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.747062 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.747442 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:09Z","lastTransitionTime":"2025-12-05T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.750417 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.750942 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.751024 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.754537 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" event={"ID":"9a27419f-8364-4ff9-a6bd-677c3fa72fef","Type":"ContainerStarted","Data":"343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.787920 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.791812 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.792694 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.813520 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.824164 4815 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.825252 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.836392 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.850083 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.850132 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.850148 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.850169 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.850185 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:09Z","lastTransitionTime":"2025-12-05T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.853471 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.872477 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.883422 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.897541 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.910109 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.930794 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.943015 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.951862 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.951893 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.951902 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.951916 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.951926 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:09Z","lastTransitionTime":"2025-12-05T09:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.954957 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.967819 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:09 crc kubenswrapper[4815]: I1205 09:07:09.985572 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.002436 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.015426 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.024030 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.036451 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.046073 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.053422 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.053470 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.053511 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.053533 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.053549 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:10Z","lastTransitionTime":"2025-12-05T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.061385 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.071261 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.083548 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.093445 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.103092 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.111297 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.120233 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.129667 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.146726 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.156036 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.156060 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.156070 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.156085 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.156095 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:10Z","lastTransitionTime":"2025-12-05T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.258014 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.258043 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.258050 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.258062 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.258071 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:10Z","lastTransitionTime":"2025-12-05T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.361463 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.361527 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.361556 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.361573 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.361585 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:10Z","lastTransitionTime":"2025-12-05T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.465195 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.465250 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.465271 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.465296 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.465314 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:10Z","lastTransitionTime":"2025-12-05T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.568185 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.568221 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.568230 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.568244 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.568253 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:10Z","lastTransitionTime":"2025-12-05T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.670698 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.670732 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.670741 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.670757 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.670766 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:10Z","lastTransitionTime":"2025-12-05T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.756707 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.773079 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.773112 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.773120 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.773133 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.773144 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:10Z","lastTransitionTime":"2025-12-05T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.875246 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.875282 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.875290 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.875303 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.875313 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:10Z","lastTransitionTime":"2025-12-05T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.977580 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.977602 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.977609 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.977622 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:10 crc kubenswrapper[4815]: I1205 09:07:10.977630 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:10Z","lastTransitionTime":"2025-12-05T09:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.079936 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.079955 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.079962 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.079973 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.079982 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:11Z","lastTransitionTime":"2025-12-05T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.183262 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.183314 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.183338 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.183370 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.183391 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:11Z","lastTransitionTime":"2025-12-05T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.284960 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.284988 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.284996 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.285007 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.285015 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:11Z","lastTransitionTime":"2025-12-05T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.386756 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.386781 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.386788 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.386800 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.386809 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:11Z","lastTransitionTime":"2025-12-05T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.418651 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.418731 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:11 crc kubenswrapper[4815]: E1205 09:07:11.418769 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.418802 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:11 crc kubenswrapper[4815]: E1205 09:07:11.418884 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:11 crc kubenswrapper[4815]: E1205 09:07:11.419009 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.438203 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.456552 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.469000 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.483265 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.489449 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.489509 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.489522 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.489540 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.489551 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:11Z","lastTransitionTime":"2025-12-05T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.496886 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.511771 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.523566 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.541726 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.552949 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.567112 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.581386 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.591824 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.591857 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.591867 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.591881 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.591891 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:11Z","lastTransitionTime":"2025-12-05T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.594775 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.606288 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.622429 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.694334 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.694371 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.694381 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.694398 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.694410 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:11Z","lastTransitionTime":"2025-12-05T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.760723 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/0.log" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.763500 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97" exitCode=1 Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.763546 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97"} Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.764216 4815 scope.go:117] "RemoveContainer" containerID="9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.780098 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.793355 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.797508 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.797545 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.797556 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.797570 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.797583 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:11Z","lastTransitionTime":"2025-12-05T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.810173 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.820542 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.833549 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.845089 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.859618 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.872418 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.893415 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:11Z\\\",\\\"message\\\":\\\"ng reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 09:07:11.070603 6045 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 09:07:11.070651 6045 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 09:07:11.070791 6045 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 09:07:11.071749 6045 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 09:07:11.071769 6045 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 09:07:11.071782 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 09:07:11.071804 6045 factory.go:656] Stopping watch factory\\\\nI1205 09:07:11.071821 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 09:07:11.071829 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 09:07:11.071836 6045 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.903592 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.903729 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.903855 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.903985 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.904068 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:11Z","lastTransitionTime":"2025-12-05T09:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.909904 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.920313 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.936371 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.954042 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:11 crc kubenswrapper[4815]: I1205 09:07:11.979397 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.017092 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.017126 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.017137 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.017155 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.017168 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:12Z","lastTransitionTime":"2025-12-05T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.120147 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.120190 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.120203 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.120219 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.120231 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:12Z","lastTransitionTime":"2025-12-05T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.222777 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.222810 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.222821 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.222836 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.222846 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:12Z","lastTransitionTime":"2025-12-05T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.344344 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.344380 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.344390 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.344404 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.344413 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:12Z","lastTransitionTime":"2025-12-05T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.446818 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.446849 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.446858 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.446871 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.446880 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:12Z","lastTransitionTime":"2025-12-05T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.548602 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.548634 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.548644 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.548657 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.548667 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:12Z","lastTransitionTime":"2025-12-05T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.650761 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.650798 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.650810 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.650828 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.650840 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:12Z","lastTransitionTime":"2025-12-05T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.753889 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.754219 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.754227 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.754245 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.754257 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:12Z","lastTransitionTime":"2025-12-05T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.768039 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/0.log" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.770362 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51"} Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.770467 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.787804 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.800923 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.816823 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.828476 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.844325 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.856934 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.856975 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.856986 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.857011 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.857023 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:12Z","lastTransitionTime":"2025-12-05T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.859615 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.875853 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.894673 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.912336 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:11Z\\\",\\\"message\\\":\\\"ng reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 09:07:11.070603 6045 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 09:07:11.070651 6045 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 09:07:11.070791 6045 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 09:07:11.071749 6045 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 09:07:11.071769 6045 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 09:07:11.071782 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 09:07:11.071804 6045 factory.go:656] Stopping watch factory\\\\nI1205 09:07:11.071821 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 09:07:11.071829 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 09:07:11.071836 6045 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.922721 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.935306 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.944198 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.958838 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.958869 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.958876 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.958888 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.958896 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:12Z","lastTransitionTime":"2025-12-05T09:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.970756 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:12 crc kubenswrapper[4815]: I1205 09:07:12.985452 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:12Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.061513 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.061551 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.061559 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.061574 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.061583 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:13Z","lastTransitionTime":"2025-12-05T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.165963 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.166021 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.166041 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.166070 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.166094 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:13Z","lastTransitionTime":"2025-12-05T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.268641 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.268909 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.269006 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.269137 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.269225 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:13Z","lastTransitionTime":"2025-12-05T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.373591 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.373666 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.373690 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.373912 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.373933 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:13Z","lastTransitionTime":"2025-12-05T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.417917 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.417955 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.418030 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:13 crc kubenswrapper[4815]: E1205 09:07:13.418074 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:13 crc kubenswrapper[4815]: E1205 09:07:13.418180 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:13 crc kubenswrapper[4815]: E1205 09:07:13.418288 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.477374 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.477409 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.477418 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.477432 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.477441 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:13Z","lastTransitionTime":"2025-12-05T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.579394 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.579434 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.579447 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.579464 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.579475 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:13Z","lastTransitionTime":"2025-12-05T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.681978 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.682016 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.682026 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.682040 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.682049 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:13Z","lastTransitionTime":"2025-12-05T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.776407 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/1.log" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.777460 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/0.log" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.782446 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51" exitCode=1 Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.782560 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51"} Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.782623 4815 scope.go:117] "RemoveContainer" containerID="9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.783576 4815 scope.go:117] "RemoveContainer" containerID="3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51" Dec 05 09:07:13 crc kubenswrapper[4815]: E1205 09:07:13.783837 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.784409 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.784465 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.784481 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.784524 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.784541 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:13Z","lastTransitionTime":"2025-12-05T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.802829 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.818431 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.830302 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.849158 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.864367 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.876789 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.887568 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.887627 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.887640 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.887654 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.887664 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:13Z","lastTransitionTime":"2025-12-05T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.889402 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.908117 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:11Z\\\",\\\"message\\\":\\\"ng reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 09:07:11.070603 6045 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 09:07:11.070651 6045 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 09:07:11.070791 6045 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 09:07:11.071749 6045 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 09:07:11.071769 6045 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 09:07:11.071782 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 09:07:11.071804 6045 factory.go:656] Stopping watch factory\\\\nI1205 09:07:11.071821 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 09:07:11.071829 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 09:07:11.071836 6045 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:13Z\\\",\\\"message\\\":\\\"Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:07:12.723214 6171 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 09:07:12.723235 6171 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1205 09:07:12.723240 6171 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.919670 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.934352 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.945607 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.969819 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.983603 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.990320 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.990369 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.990380 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.990400 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.990413 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:13Z","lastTransitionTime":"2025-12-05T09:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:13 crc kubenswrapper[4815]: I1205 09:07:13.993361 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:13Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.092291 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.092334 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.092347 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.092363 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.092374 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.194603 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.194629 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.194636 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.194650 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.194659 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.204407 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5"] Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.204884 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.208305 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.209410 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.227381 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.241100 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.253709 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.267251 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.280169 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.292217 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.296925 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.296961 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.296969 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.296982 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.296991 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.304953 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.316373 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.335216 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9aea640c421a8b85d1e793019f4a43af55f233190de59eadffaea00de04e8d97\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:11Z\\\",\\\"message\\\":\\\"ng reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 09:07:11.070603 6045 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 09:07:11.070651 6045 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1205 09:07:11.070791 6045 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1205 09:07:11.071749 6045 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1205 09:07:11.071769 6045 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1205 09:07:11.071782 6045 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1205 09:07:11.071804 6045 factory.go:656] Stopping watch factory\\\\nI1205 09:07:11.071821 6045 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1205 09:07:11.071829 6045 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1205 09:07:11.071836 6045 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:13Z\\\",\\\"message\\\":\\\"Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:07:12.723214 6171 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 09:07:12.723235 6171 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1205 09:07:12.723240 6171 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.345478 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.362605 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.372471 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsxf9\" (UniqueName: \"kubernetes.io/projected/bd567f9e-eb99-4085-97ee-63a58de73556-kube-api-access-hsxf9\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.372518 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd567f9e-eb99-4085-97ee-63a58de73556-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.372549 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd567f9e-eb99-4085-97ee-63a58de73556-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.372567 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd567f9e-eb99-4085-97ee-63a58de73556-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.377050 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.387292 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.398674 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.399229 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.399272 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.399282 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.399299 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.399318 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.408030 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.418378 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.418426 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.418439 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.418456 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.418468 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: E1205 09:07:14.429598 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.432181 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.432228 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.432238 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.432253 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.432266 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: E1205 09:07:14.443048 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.446165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.446195 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.446204 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.446217 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.446225 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: E1205 09:07:14.456312 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.459588 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.459637 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.459653 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.459672 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.459688 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: E1205 09:07:14.470445 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.472977 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd567f9e-eb99-4085-97ee-63a58de73556-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.473012 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsxf9\" (UniqueName: \"kubernetes.io/projected/bd567f9e-eb99-4085-97ee-63a58de73556-kube-api-access-hsxf9\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.473055 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd567f9e-eb99-4085-97ee-63a58de73556-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.473081 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd567f9e-eb99-4085-97ee-63a58de73556-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.474116 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/bd567f9e-eb99-4085-97ee-63a58de73556-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.474148 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/bd567f9e-eb99-4085-97ee-63a58de73556-env-overrides\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.475134 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.475171 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.475179 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.475193 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.475202 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.486903 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/bd567f9e-eb99-4085-97ee-63a58de73556-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: E1205 09:07:14.489241 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: E1205 09:07:14.489391 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.495933 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsxf9\" (UniqueName: \"kubernetes.io/projected/bd567f9e-eb99-4085-97ee-63a58de73556-kube-api-access-hsxf9\") pod \"ovnkube-control-plane-749d76644c-xd2x5\" (UID: \"bd567f9e-eb99-4085-97ee-63a58de73556\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.501775 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.501815 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.501823 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.501837 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.501847 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.519951 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" Dec 05 09:07:14 crc kubenswrapper[4815]: W1205 09:07:14.531787 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd567f9e_eb99_4085_97ee_63a58de73556.slice/crio-115d30b4067fcd7eefa3a853a03b9e06d9bc6a9c1ee5b71010fd72ea0320454b WatchSource:0}: Error finding container 115d30b4067fcd7eefa3a853a03b9e06d9bc6a9c1ee5b71010fd72ea0320454b: Status 404 returned error can't find the container with id 115d30b4067fcd7eefa3a853a03b9e06d9bc6a9c1ee5b71010fd72ea0320454b Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.603815 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.603882 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.603896 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.603912 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.603923 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.712510 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.712548 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.712559 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.712575 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.712587 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.789712 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/1.log" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.796929 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" event={"ID":"bd567f9e-eb99-4085-97ee-63a58de73556","Type":"ContainerStarted","Data":"115d30b4067fcd7eefa3a853a03b9e06d9bc6a9c1ee5b71010fd72ea0320454b"} Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.797686 4815 scope.go:117] "RemoveContainer" containerID="3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51" Dec 05 09:07:14 crc kubenswrapper[4815]: E1205 09:07:14.797987 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.815987 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.816033 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.816045 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.816061 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.816073 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.818024 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.838713 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:13Z\\\",\\\"message\\\":\\\"Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:07:12.723214 6171 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 09:07:12.723235 6171 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1205 09:07:12.723240 6171 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.850153 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.860926 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.872166 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.895696 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.921444 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.921472 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.921479 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.921511 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.921520 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:14Z","lastTransitionTime":"2025-12-05T09:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.922928 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.944214 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.955189 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.970282 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.979533 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:14 crc kubenswrapper[4815]: I1205 09:07:14.990440 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:14Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.005813 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.023480 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.023743 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.023782 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.023798 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.023818 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.023833 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:15Z","lastTransitionTime":"2025-12-05T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.040965 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.126432 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.126456 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.126464 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.126477 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.126504 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:15Z","lastTransitionTime":"2025-12-05T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.228896 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.228948 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.228967 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.228993 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.229014 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:15Z","lastTransitionTime":"2025-12-05T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.294211 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-sslm2"] Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.294668 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:15 crc kubenswrapper[4815]: E1205 09:07:15.294725 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.319276 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.331905 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.331950 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.331962 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.331977 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.331989 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:15Z","lastTransitionTime":"2025-12-05T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.340333 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.350318 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.360377 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.372468 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.384560 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.395233 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.411532 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.417958 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:15 crc kubenswrapper[4815]: E1205 09:07:15.418107 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.418589 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:15 crc kubenswrapper[4815]: E1205 09:07:15.418690 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.418850 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:15 crc kubenswrapper[4815]: E1205 09:07:15.418947 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.432292 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.433867 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.433894 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.433904 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.433920 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.433931 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:15Z","lastTransitionTime":"2025-12-05T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.450588 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:13Z\\\",\\\"message\\\":\\\"Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:07:12.723214 6171 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 09:07:12.723235 6171 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1205 09:07:12.723240 6171 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.461906 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.475907 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.483278 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.483319 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmrzd\" (UniqueName: \"kubernetes.io/projected/010248a6-d07e-4e6b-97c8-bda1925d6adb-kube-api-access-vmrzd\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.489079 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.500914 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.512199 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.530155 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.536403 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.536437 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.536448 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.536520 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.536533 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:15Z","lastTransitionTime":"2025-12-05T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.584635 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.584685 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmrzd\" (UniqueName: \"kubernetes.io/projected/010248a6-d07e-4e6b-97c8-bda1925d6adb-kube-api-access-vmrzd\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:15 crc kubenswrapper[4815]: E1205 09:07:15.584801 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:15 crc kubenswrapper[4815]: E1205 09:07:15.584890 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs podName:010248a6-d07e-4e6b-97c8-bda1925d6adb nodeName:}" failed. No retries permitted until 2025-12-05 09:07:16.084870927 +0000 UTC m=+34.963477754 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs") pod "network-metrics-daemon-sslm2" (UID: "010248a6-d07e-4e6b-97c8-bda1925d6adb") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.606303 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmrzd\" (UniqueName: \"kubernetes.io/projected/010248a6-d07e-4e6b-97c8-bda1925d6adb-kube-api-access-vmrzd\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.639202 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.639239 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.639250 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.639267 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.639279 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:15Z","lastTransitionTime":"2025-12-05T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.741317 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.741355 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.741364 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.741380 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.741392 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:15Z","lastTransitionTime":"2025-12-05T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.802029 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" event={"ID":"bd567f9e-eb99-4085-97ee-63a58de73556","Type":"ContainerStarted","Data":"6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.802107 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" event={"ID":"bd567f9e-eb99-4085-97ee-63a58de73556","Type":"ContainerStarted","Data":"f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.816595 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.827506 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.844118 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.844171 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.844184 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.844203 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.844216 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:15Z","lastTransitionTime":"2025-12-05T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.846875 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.859760 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.873211 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.889015 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.910677 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.925831 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.941934 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.946759 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.946812 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.946829 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.946853 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.946870 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:15Z","lastTransitionTime":"2025-12-05T09:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.957879 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.980576 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:15 crc kubenswrapper[4815]: I1205 09:07:15.993904 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:15Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.026750 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:13Z\\\",\\\"message\\\":\\\"Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:07:12.723214 6171 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 09:07:12.723235 6171 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1205 09:07:12.723240 6171 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.042816 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.051106 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.051534 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.051631 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.051731 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.051900 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:16Z","lastTransitionTime":"2025-12-05T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.057448 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.070608 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:16Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.090636 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:16 crc kubenswrapper[4815]: E1205 09:07:16.090914 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:16 crc kubenswrapper[4815]: E1205 09:07:16.091007 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs podName:010248a6-d07e-4e6b-97c8-bda1925d6adb nodeName:}" failed. No retries permitted until 2025-12-05 09:07:17.090976746 +0000 UTC m=+35.969583623 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs") pod "network-metrics-daemon-sslm2" (UID: "010248a6-d07e-4e6b-97c8-bda1925d6adb") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.156317 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.156365 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.156376 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.156394 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.156405 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:16Z","lastTransitionTime":"2025-12-05T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.258950 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.258977 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.258985 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.258998 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.259006 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:16Z","lastTransitionTime":"2025-12-05T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.361713 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.361769 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.361787 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.361810 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.361826 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:16Z","lastTransitionTime":"2025-12-05T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.464238 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.464286 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.464295 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.464308 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.464318 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:16Z","lastTransitionTime":"2025-12-05T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.569807 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.569843 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.569852 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.569871 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.569887 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:16Z","lastTransitionTime":"2025-12-05T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.672257 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.672297 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.672308 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.672327 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.672339 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:16Z","lastTransitionTime":"2025-12-05T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.775132 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.775176 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.775192 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.775214 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.775230 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:16Z","lastTransitionTime":"2025-12-05T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.877723 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.877788 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.877812 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.877842 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.877864 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:16Z","lastTransitionTime":"2025-12-05T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.980213 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.980264 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.980278 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.980297 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:16 crc kubenswrapper[4815]: I1205 09:07:16.980310 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:16Z","lastTransitionTime":"2025-12-05T09:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.082943 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.082979 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.082988 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.083003 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.083013 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:17Z","lastTransitionTime":"2025-12-05T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.102375 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.102556 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.103192 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs podName:010248a6-d07e-4e6b-97c8-bda1925d6adb nodeName:}" failed. No retries permitted until 2025-12-05 09:07:19.103136275 +0000 UTC m=+37.981743112 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs") pod "network-metrics-daemon-sslm2" (UID: "010248a6-d07e-4e6b-97c8-bda1925d6adb") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.185788 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.185818 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.185827 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.185842 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.185852 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:17Z","lastTransitionTime":"2025-12-05T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.204607 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.204804 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:07:33.204784951 +0000 UTC m=+52.083391788 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.288645 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.288686 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.288698 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.288715 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.288726 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:17Z","lastTransitionTime":"2025-12-05T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.305705 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.305746 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.305796 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.305830 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.305914 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.305921 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.305954 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.305968 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.306052 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:33.306024665 +0000 UTC m=+52.184631522 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.306065 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.306081 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.306092 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.306125 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.306129 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:33.306113318 +0000 UTC m=+52.184720195 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.306209 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:33.30619183 +0000 UTC m=+52.184798717 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.306229 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:07:33.306218381 +0000 UTC m=+52.184825298 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.391176 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.391204 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.391214 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.391227 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.391235 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:17Z","lastTransitionTime":"2025-12-05T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.418553 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.418629 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.418732 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.418779 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.418569 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.418874 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.419112 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:17 crc kubenswrapper[4815]: E1205 09:07:17.419450 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.493706 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.493746 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.493758 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.493774 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.493785 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:17Z","lastTransitionTime":"2025-12-05T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.596101 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.596150 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.596166 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.596186 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.596201 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:17Z","lastTransitionTime":"2025-12-05T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.698761 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.698837 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.698858 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.698878 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.698942 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:17Z","lastTransitionTime":"2025-12-05T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.801997 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.802066 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.802077 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.802094 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.802111 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:17Z","lastTransitionTime":"2025-12-05T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.904157 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.904189 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.904198 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.904213 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:17 crc kubenswrapper[4815]: I1205 09:07:17.904222 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:17Z","lastTransitionTime":"2025-12-05T09:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.006961 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.006990 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.006997 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.007012 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.007020 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:18Z","lastTransitionTime":"2025-12-05T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.109751 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.109790 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.109800 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.109817 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.109828 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:18Z","lastTransitionTime":"2025-12-05T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.212475 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.212529 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.212537 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.212553 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.212565 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:18Z","lastTransitionTime":"2025-12-05T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.314914 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.314972 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.314985 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.315006 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.315018 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:18Z","lastTransitionTime":"2025-12-05T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.417693 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.417733 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.417743 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.417761 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.417772 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:18Z","lastTransitionTime":"2025-12-05T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.520153 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.520188 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.520198 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.520214 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.520227 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:18Z","lastTransitionTime":"2025-12-05T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.622719 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.622783 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.622808 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.622828 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.622842 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:18Z","lastTransitionTime":"2025-12-05T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.724966 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.725009 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.725020 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.725037 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.725048 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:18Z","lastTransitionTime":"2025-12-05T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.828226 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.828299 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.828341 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.828368 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.828386 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:18Z","lastTransitionTime":"2025-12-05T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.930322 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.930352 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.930362 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.930376 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:18 crc kubenswrapper[4815]: I1205 09:07:18.930386 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:18Z","lastTransitionTime":"2025-12-05T09:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.032056 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.032093 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.032103 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.032118 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.032129 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:19Z","lastTransitionTime":"2025-12-05T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.124146 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:19 crc kubenswrapper[4815]: E1205 09:07:19.124312 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:19 crc kubenswrapper[4815]: E1205 09:07:19.124399 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs podName:010248a6-d07e-4e6b-97c8-bda1925d6adb nodeName:}" failed. No retries permitted until 2025-12-05 09:07:23.124379884 +0000 UTC m=+42.002986721 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs") pod "network-metrics-daemon-sslm2" (UID: "010248a6-d07e-4e6b-97c8-bda1925d6adb") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.134913 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.134941 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.134950 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.134961 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.134971 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:19Z","lastTransitionTime":"2025-12-05T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.238057 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.238095 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.238105 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.238119 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.238128 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:19Z","lastTransitionTime":"2025-12-05T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.340991 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.341052 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.341072 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.341095 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.341117 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:19Z","lastTransitionTime":"2025-12-05T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.357263 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.380365 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.394595 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.408308 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.417895 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:19 crc kubenswrapper[4815]: E1205 09:07:19.418044 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.418116 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:19 crc kubenswrapper[4815]: E1205 09:07:19.418173 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.418509 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:19 crc kubenswrapper[4815]: E1205 09:07:19.418581 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.418738 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:19 crc kubenswrapper[4815]: E1205 09:07:19.418811 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.423591 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.437362 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.443325 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.443362 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.443371 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.443421 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.443431 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:19Z","lastTransitionTime":"2025-12-05T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.450289 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.470611 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:13Z\\\",\\\"message\\\":\\\"Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:07:12.723214 6171 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 09:07:12.723235 6171 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1205 09:07:12.723240 6171 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.483025 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.492753 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.505169 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.518451 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.532750 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.545516 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.545561 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.545558 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.545596 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.545699 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.545710 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:19Z","lastTransitionTime":"2025-12-05T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.561974 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.574679 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.584933 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:19Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.647898 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.647987 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.648006 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.648068 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.648092 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:19Z","lastTransitionTime":"2025-12-05T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.751404 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.751551 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.751571 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.751634 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.751660 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:19Z","lastTransitionTime":"2025-12-05T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.854417 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.854476 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.854526 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.854548 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.854562 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:19Z","lastTransitionTime":"2025-12-05T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.957709 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.957746 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.957757 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.957773 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:19 crc kubenswrapper[4815]: I1205 09:07:19.957784 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:19Z","lastTransitionTime":"2025-12-05T09:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.060045 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.060103 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.060117 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.060138 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.060154 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:20Z","lastTransitionTime":"2025-12-05T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.162409 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.162448 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.162458 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.162472 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.162512 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:20Z","lastTransitionTime":"2025-12-05T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.265483 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.265556 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.265566 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.265579 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.265589 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:20Z","lastTransitionTime":"2025-12-05T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.367715 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.367768 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.367791 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.367813 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.367828 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:20Z","lastTransitionTime":"2025-12-05T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.470764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.470815 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.470826 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.470842 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.470855 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:20Z","lastTransitionTime":"2025-12-05T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.573201 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.573236 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.573247 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.573261 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.573272 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:20Z","lastTransitionTime":"2025-12-05T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.675564 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.675623 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.675639 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.675661 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.675677 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:20Z","lastTransitionTime":"2025-12-05T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.778050 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.778084 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.778093 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.778106 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.778116 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:20Z","lastTransitionTime":"2025-12-05T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.880692 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.880729 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.880739 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.880754 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.880765 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:20Z","lastTransitionTime":"2025-12-05T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.983509 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.983582 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.983592 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.983606 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:20 crc kubenswrapper[4815]: I1205 09:07:20.983614 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:20Z","lastTransitionTime":"2025-12-05T09:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.086281 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.086315 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.086326 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.086341 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.086350 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:21Z","lastTransitionTime":"2025-12-05T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.189166 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.189249 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.189275 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.189305 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.189328 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:21Z","lastTransitionTime":"2025-12-05T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.292572 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.292647 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.292670 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.292703 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.292725 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:21Z","lastTransitionTime":"2025-12-05T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.396248 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.396325 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.396353 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.396383 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.396405 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:21Z","lastTransitionTime":"2025-12-05T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.417673 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:21 crc kubenswrapper[4815]: E1205 09:07:21.417895 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.418022 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.418089 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:21 crc kubenswrapper[4815]: E1205 09:07:21.418192 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.418022 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:21 crc kubenswrapper[4815]: E1205 09:07:21.418343 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:21 crc kubenswrapper[4815]: E1205 09:07:21.418600 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.433842 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.449116 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.476413 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.494905 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.498515 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.498548 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.498561 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.498576 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.498588 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:21Z","lastTransitionTime":"2025-12-05T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.514525 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.527099 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.553128 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:13Z\\\",\\\"message\\\":\\\"Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:07:12.723214 6171 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 09:07:12.723235 6171 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1205 09:07:12.723240 6171 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.564760 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.573986 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.585385 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.594638 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.600897 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.600952 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.600962 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.600976 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.600984 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:21Z","lastTransitionTime":"2025-12-05T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.606380 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.616156 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.632282 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.643865 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.652546 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.703181 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.703239 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.703257 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.703280 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.703298 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:21Z","lastTransitionTime":"2025-12-05T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.807012 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.807072 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.807089 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.807117 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.807134 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:21Z","lastTransitionTime":"2025-12-05T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.909703 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.909773 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.909786 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.909802 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:21 crc kubenswrapper[4815]: I1205 09:07:21.909814 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:21Z","lastTransitionTime":"2025-12-05T09:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.015214 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.015328 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.015356 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.015384 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.015406 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:22Z","lastTransitionTime":"2025-12-05T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.118366 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.118447 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.118467 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.118525 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.118546 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:22Z","lastTransitionTime":"2025-12-05T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.220777 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.220853 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.220876 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.220905 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.220928 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:22Z","lastTransitionTime":"2025-12-05T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.324013 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.324069 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.324085 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.324109 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.324125 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:22Z","lastTransitionTime":"2025-12-05T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.427840 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.427912 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.427935 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.427962 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.427984 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:22Z","lastTransitionTime":"2025-12-05T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.530798 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.530864 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.530886 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.530916 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.530935 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:22Z","lastTransitionTime":"2025-12-05T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.634539 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.634602 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.634620 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.634641 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.634653 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:22Z","lastTransitionTime":"2025-12-05T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.737155 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.737215 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.737232 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.737257 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.737278 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:22Z","lastTransitionTime":"2025-12-05T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.840637 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.840697 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.840715 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.840751 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.840772 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:22Z","lastTransitionTime":"2025-12-05T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.943928 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.943987 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.944002 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.944018 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:22 crc kubenswrapper[4815]: I1205 09:07:22.944029 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:22Z","lastTransitionTime":"2025-12-05T09:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.046456 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.046523 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.046533 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.046548 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.046557 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:23Z","lastTransitionTime":"2025-12-05T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.149065 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.149102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.149113 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.149129 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.149139 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:23Z","lastTransitionTime":"2025-12-05T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.159881 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:23 crc kubenswrapper[4815]: E1205 09:07:23.160071 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:23 crc kubenswrapper[4815]: E1205 09:07:23.160150 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs podName:010248a6-d07e-4e6b-97c8-bda1925d6adb nodeName:}" failed. No retries permitted until 2025-12-05 09:07:31.160128523 +0000 UTC m=+50.038735380 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs") pod "network-metrics-daemon-sslm2" (UID: "010248a6-d07e-4e6b-97c8-bda1925d6adb") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.252082 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.252143 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.252160 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.252220 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.252239 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:23Z","lastTransitionTime":"2025-12-05T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.354992 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.355031 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.355042 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.355056 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.355067 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:23Z","lastTransitionTime":"2025-12-05T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.365714 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.367050 4815 scope.go:117] "RemoveContainer" containerID="3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51" Dec 05 09:07:23 crc kubenswrapper[4815]: E1205 09:07:23.367366 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.418289 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.418338 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:23 crc kubenswrapper[4815]: E1205 09:07:23.418473 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.418634 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.418648 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:23 crc kubenswrapper[4815]: E1205 09:07:23.418752 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:23 crc kubenswrapper[4815]: E1205 09:07:23.418843 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:23 crc kubenswrapper[4815]: E1205 09:07:23.418902 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.458256 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.458302 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.458313 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.458331 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.458396 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:23Z","lastTransitionTime":"2025-12-05T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.561216 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.561263 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.561273 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.561313 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.561332 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:23Z","lastTransitionTime":"2025-12-05T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.664426 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.664468 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.664479 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.664517 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.664557 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:23Z","lastTransitionTime":"2025-12-05T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.767192 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.767251 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.767266 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.767287 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.767302 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:23Z","lastTransitionTime":"2025-12-05T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.869421 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.869549 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.869575 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.869603 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.869628 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:23Z","lastTransitionTime":"2025-12-05T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.972843 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.972888 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.972900 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.972919 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:23 crc kubenswrapper[4815]: I1205 09:07:23.972932 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:23Z","lastTransitionTime":"2025-12-05T09:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.075588 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.075629 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.075639 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.075657 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.075672 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.178766 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.178835 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.178847 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.178871 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.178885 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.281244 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.281296 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.281307 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.281327 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.281344 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.384686 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.384736 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.384759 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.384790 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.384805 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.487912 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.487967 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.487979 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.487998 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.488009 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.590304 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.590355 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.590372 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.590394 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.590410 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.693307 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.693356 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.693366 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.693381 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.693392 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.795934 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.796017 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.796046 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.796116 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.796138 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.819311 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.819361 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.819378 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.819398 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.819412 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: E1205 09:07:24.830930 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:24Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.834536 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.834579 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.834591 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.834609 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.834621 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: E1205 09:07:24.847373 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:24Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.851794 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.851834 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.851849 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.851867 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.851878 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: E1205 09:07:24.869262 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:24Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.872311 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.872362 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.872374 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.872392 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.872404 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: E1205 09:07:24.884896 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:24Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.888036 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.888081 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.888096 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.888115 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.888126 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:24 crc kubenswrapper[4815]: E1205 09:07:24.899395 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:24Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:24 crc kubenswrapper[4815]: E1205 09:07:24.899587 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.901088 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.901123 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.901145 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.901162 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:24 crc kubenswrapper[4815]: I1205 09:07:24.901174 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:24Z","lastTransitionTime":"2025-12-05T09:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.003270 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.003320 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.003332 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.003352 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.003384 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:25Z","lastTransitionTime":"2025-12-05T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.106141 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.106182 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.106193 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.106211 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.106222 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:25Z","lastTransitionTime":"2025-12-05T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.208486 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.208580 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.208606 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.208632 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.208654 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:25Z","lastTransitionTime":"2025-12-05T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.311998 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.312058 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.312076 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.312098 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.312161 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:25Z","lastTransitionTime":"2025-12-05T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.414884 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.414929 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.414941 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.414957 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.414968 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:25Z","lastTransitionTime":"2025-12-05T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.418735 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:25 crc kubenswrapper[4815]: E1205 09:07:25.418979 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.419116 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.419161 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.419159 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:25 crc kubenswrapper[4815]: E1205 09:07:25.419227 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:25 crc kubenswrapper[4815]: E1205 09:07:25.419357 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:25 crc kubenswrapper[4815]: E1205 09:07:25.419472 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.518036 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.518082 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.518098 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.518120 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.518136 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:25Z","lastTransitionTime":"2025-12-05T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.621368 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.621403 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.621413 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.621429 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.621440 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:25Z","lastTransitionTime":"2025-12-05T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.724092 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.724130 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.724140 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.724156 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.724167 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:25Z","lastTransitionTime":"2025-12-05T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.825831 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.825870 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.825880 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.825896 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.825906 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:25Z","lastTransitionTime":"2025-12-05T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.928956 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.929017 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.929041 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.929077 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:25 crc kubenswrapper[4815]: I1205 09:07:25.929101 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:25Z","lastTransitionTime":"2025-12-05T09:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.032150 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.032225 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.032249 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.032279 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.032301 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:26Z","lastTransitionTime":"2025-12-05T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.134662 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.134737 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.134756 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.134775 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.134788 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:26Z","lastTransitionTime":"2025-12-05T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.237447 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.237576 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.237600 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.237642 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.237665 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:26Z","lastTransitionTime":"2025-12-05T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.341270 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.341344 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.341365 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.341393 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.341414 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:26Z","lastTransitionTime":"2025-12-05T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.444922 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.444975 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.444992 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.445016 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.445033 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:26Z","lastTransitionTime":"2025-12-05T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.548243 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.548310 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.548346 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.548382 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.548405 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:26Z","lastTransitionTime":"2025-12-05T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.651426 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.651530 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.651554 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.651587 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.651609 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:26Z","lastTransitionTime":"2025-12-05T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.754787 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.755265 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.755353 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.755433 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.755553 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:26Z","lastTransitionTime":"2025-12-05T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.858188 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.858250 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.858274 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.858321 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.858335 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:26Z","lastTransitionTime":"2025-12-05T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.961948 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.962008 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.962023 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.962045 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:26 crc kubenswrapper[4815]: I1205 09:07:26.962062 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:26Z","lastTransitionTime":"2025-12-05T09:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.065594 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.065654 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.065671 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.065694 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.065714 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:27Z","lastTransitionTime":"2025-12-05T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.168682 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.168729 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.168741 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.168757 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.168769 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:27Z","lastTransitionTime":"2025-12-05T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.270902 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.270933 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.270940 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.270953 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.270962 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:27Z","lastTransitionTime":"2025-12-05T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.373743 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.373779 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.373788 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.373802 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.373812 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:27Z","lastTransitionTime":"2025-12-05T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.418159 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.418264 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.418190 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.418159 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:27 crc kubenswrapper[4815]: E1205 09:07:27.418452 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:27 crc kubenswrapper[4815]: E1205 09:07:27.418551 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:27 crc kubenswrapper[4815]: E1205 09:07:27.418668 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:27 crc kubenswrapper[4815]: E1205 09:07:27.418793 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.476734 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.476781 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.476793 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.476812 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.476825 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:27Z","lastTransitionTime":"2025-12-05T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.579021 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.579066 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.579078 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.579093 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.579102 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:27Z","lastTransitionTime":"2025-12-05T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.681856 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.681910 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.681928 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.681951 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.681968 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:27Z","lastTransitionTime":"2025-12-05T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.784410 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.784782 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.784999 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.785270 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.785421 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:27Z","lastTransitionTime":"2025-12-05T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.888302 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.888365 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.888383 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.888406 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.888423 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:27Z","lastTransitionTime":"2025-12-05T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.991803 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.991907 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.991931 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.991956 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:27 crc kubenswrapper[4815]: I1205 09:07:27.991976 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:27Z","lastTransitionTime":"2025-12-05T09:07:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.095301 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.095410 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.095433 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.095465 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.095494 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:28Z","lastTransitionTime":"2025-12-05T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.199892 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.199967 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.199989 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.200017 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.200038 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:28Z","lastTransitionTime":"2025-12-05T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.305978 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.306047 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.306066 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.306089 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.306107 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:28Z","lastTransitionTime":"2025-12-05T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.408999 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.409317 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.409613 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.409809 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.410070 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:28Z","lastTransitionTime":"2025-12-05T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.513060 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.513408 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.513589 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.513764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.513886 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:28Z","lastTransitionTime":"2025-12-05T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.616899 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.616946 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.616957 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.616975 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.616989 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:28Z","lastTransitionTime":"2025-12-05T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.719819 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.719857 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.719865 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.719881 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.719891 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:28Z","lastTransitionTime":"2025-12-05T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.822803 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.822857 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.822872 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.822893 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.822913 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:28Z","lastTransitionTime":"2025-12-05T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.925767 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.925805 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.925814 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.925828 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:28 crc kubenswrapper[4815]: I1205 09:07:28.925839 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:28Z","lastTransitionTime":"2025-12-05T09:07:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.028735 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.028767 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.028777 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.028792 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.028803 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:29Z","lastTransitionTime":"2025-12-05T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.131268 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.131311 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.131321 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.131337 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.131348 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:29Z","lastTransitionTime":"2025-12-05T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.233894 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.233930 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.233939 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.233954 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.233963 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:29Z","lastTransitionTime":"2025-12-05T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.340472 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.340890 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.340979 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.341073 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.341188 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:29Z","lastTransitionTime":"2025-12-05T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.418082 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.418134 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.418428 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.418585 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:29 crc kubenswrapper[4815]: E1205 09:07:29.418613 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:29 crc kubenswrapper[4815]: E1205 09:07:29.418722 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:29 crc kubenswrapper[4815]: E1205 09:07:29.418864 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:29 crc kubenswrapper[4815]: E1205 09:07:29.419003 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.443701 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.443768 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.443781 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.443799 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.443811 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:29Z","lastTransitionTime":"2025-12-05T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.546942 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.547023 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.547049 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.547079 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.547102 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:29Z","lastTransitionTime":"2025-12-05T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.650419 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.650483 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.650545 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.650574 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.650598 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:29Z","lastTransitionTime":"2025-12-05T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.753836 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.753876 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.753886 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.753902 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.753914 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:29Z","lastTransitionTime":"2025-12-05T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.855710 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.855743 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.855754 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.855777 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.855788 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:29Z","lastTransitionTime":"2025-12-05T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.958994 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.959066 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.959083 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.959107 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:29 crc kubenswrapper[4815]: I1205 09:07:29.959124 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:29Z","lastTransitionTime":"2025-12-05T09:07:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.061473 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.061764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.061828 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.061892 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.061963 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:30Z","lastTransitionTime":"2025-12-05T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.164911 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.165188 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.165256 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.165319 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.165392 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:30Z","lastTransitionTime":"2025-12-05T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.268202 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.268252 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.268262 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.268276 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.268285 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:30Z","lastTransitionTime":"2025-12-05T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.371172 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.371587 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.371694 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.371773 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.371848 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:30Z","lastTransitionTime":"2025-12-05T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.474844 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.474877 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.474908 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.474922 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.474932 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:30Z","lastTransitionTime":"2025-12-05T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.524366 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.533552 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.544035 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.552546 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.564267 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.577800 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.577844 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.577882 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.577928 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.577968 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:30Z","lastTransitionTime":"2025-12-05T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.579315 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.590614 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.606141 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.616304 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.626013 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.636030 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.651342 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:13Z\\\",\\\"message\\\":\\\"Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:07:12.723214 6171 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 09:07:12.723235 6171 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1205 09:07:12.723240 6171 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.670103 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.680247 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.680274 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.680283 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.680296 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.680307 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:30Z","lastTransitionTime":"2025-12-05T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.685184 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.697750 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.710480 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.719727 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.749431 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:30Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.782882 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.782940 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.782958 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.782984 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.783006 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:30Z","lastTransitionTime":"2025-12-05T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.885046 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.885086 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.885097 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.885110 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.885120 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:30Z","lastTransitionTime":"2025-12-05T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.987857 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.987906 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.987920 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.987937 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:30 crc kubenswrapper[4815]: I1205 09:07:30.987950 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:30Z","lastTransitionTime":"2025-12-05T09:07:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.089811 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.089846 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.089855 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.089869 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.089879 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:31Z","lastTransitionTime":"2025-12-05T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.191975 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.192018 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.192028 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.192045 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.192057 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:31Z","lastTransitionTime":"2025-12-05T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.243627 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:31 crc kubenswrapper[4815]: E1205 09:07:31.243765 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:31 crc kubenswrapper[4815]: E1205 09:07:31.243840 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs podName:010248a6-d07e-4e6b-97c8-bda1925d6adb nodeName:}" failed. No retries permitted until 2025-12-05 09:07:47.243819712 +0000 UTC m=+66.122426549 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs") pod "network-metrics-daemon-sslm2" (UID: "010248a6-d07e-4e6b-97c8-bda1925d6adb") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.295603 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.295660 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.295670 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.295691 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.295703 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:31Z","lastTransitionTime":"2025-12-05T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.398948 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.399001 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.399017 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.399040 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.399056 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:31Z","lastTransitionTime":"2025-12-05T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.418730 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.418809 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:31 crc kubenswrapper[4815]: E1205 09:07:31.418884 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:31 crc kubenswrapper[4815]: E1205 09:07:31.419053 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.419200 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.419329 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:31 crc kubenswrapper[4815]: E1205 09:07:31.419411 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:31 crc kubenswrapper[4815]: E1205 09:07:31.419576 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.434996 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.450037 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.464451 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.479748 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.498193 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.502563 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.502620 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.502635 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.502659 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.502672 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:31Z","lastTransitionTime":"2025-12-05T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.515088 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.528418 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.539877 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.558479 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.576499 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.603368 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:13Z\\\",\\\"message\\\":\\\"Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:07:12.723214 6171 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 09:07:12.723235 6171 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1205 09:07:12.723240 6171 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.605368 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.605409 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.605418 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.605438 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.605448 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:31Z","lastTransitionTime":"2025-12-05T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.614598 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.624928 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.638402 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.661257 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.672750 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.693637 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:31Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.708435 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.708481 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.708499 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.708553 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.708568 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:31Z","lastTransitionTime":"2025-12-05T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.811472 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.811561 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.811573 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.811615 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.811627 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:31Z","lastTransitionTime":"2025-12-05T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.915022 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.915095 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.915117 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.915143 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:31 crc kubenswrapper[4815]: I1205 09:07:31.915160 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:31Z","lastTransitionTime":"2025-12-05T09:07:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.018474 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.018564 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.018605 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.018632 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.018654 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:32Z","lastTransitionTime":"2025-12-05T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.121610 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.121710 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.121729 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.121752 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.121804 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:32Z","lastTransitionTime":"2025-12-05T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.225623 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.225664 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.225793 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.225830 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.225854 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:32Z","lastTransitionTime":"2025-12-05T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.328713 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.328759 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.328771 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.328788 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.328801 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:32Z","lastTransitionTime":"2025-12-05T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.431745 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.431779 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.431787 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.431800 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.431862 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:32Z","lastTransitionTime":"2025-12-05T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.535544 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.535613 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.535635 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.535659 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.535673 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:32Z","lastTransitionTime":"2025-12-05T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.638808 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.638848 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.638857 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.638870 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.638879 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:32Z","lastTransitionTime":"2025-12-05T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.783599 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.783635 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.783643 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.783657 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.783666 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:32Z","lastTransitionTime":"2025-12-05T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.887031 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.887060 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.887068 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.887102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.887111 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:32Z","lastTransitionTime":"2025-12-05T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.989185 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.989266 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.989284 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.989352 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:32 crc kubenswrapper[4815]: I1205 09:07:32.989370 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:32Z","lastTransitionTime":"2025-12-05T09:07:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.093033 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.093113 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.093132 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.093175 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.093200 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:33Z","lastTransitionTime":"2025-12-05T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.196131 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.196182 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.196193 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.196210 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.196222 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:33Z","lastTransitionTime":"2025-12-05T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.290673 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.290968 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:08:05.290939302 +0000 UTC m=+84.169546149 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.298599 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.298690 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.298710 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.298764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.298778 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:33Z","lastTransitionTime":"2025-12-05T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.392161 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.392212 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.392257 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.392290 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392302 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392426 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392428 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392442 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392448 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392460 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392347 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392566 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392522 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 09:08:05.392474665 +0000 UTC m=+84.271081512 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392663 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:08:05.392633789 +0000 UTC m=+84.271240666 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392689 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 09:08:05.39267719 +0000 UTC m=+84.271284067 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.392711 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:08:05.392699551 +0000 UTC m=+84.271306428 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.402071 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.402119 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.402136 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.402157 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.402173 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:33Z","lastTransitionTime":"2025-12-05T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.418602 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.418632 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.418666 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.418617 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.418752 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.418829 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.418941 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:33 crc kubenswrapper[4815]: E1205 09:07:33.418991 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.504824 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.504864 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.504875 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.504891 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.504903 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:33Z","lastTransitionTime":"2025-12-05T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.607622 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.607668 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.607679 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.607695 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.607706 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:33Z","lastTransitionTime":"2025-12-05T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.710180 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.710233 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.710249 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.710272 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.710290 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:33Z","lastTransitionTime":"2025-12-05T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.812574 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.812634 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.812647 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.812663 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.812675 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:33Z","lastTransitionTime":"2025-12-05T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.915951 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.915998 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.916011 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.916029 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:33 crc kubenswrapper[4815]: I1205 09:07:33.916042 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:33Z","lastTransitionTime":"2025-12-05T09:07:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.018560 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.018610 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.018626 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.018642 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.018653 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:34Z","lastTransitionTime":"2025-12-05T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.121899 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.121946 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.121957 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.121972 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.121984 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:34Z","lastTransitionTime":"2025-12-05T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.224122 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.224165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.224287 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.224301 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.224310 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:34Z","lastTransitionTime":"2025-12-05T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.325878 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.325905 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.325912 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.325925 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.325933 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:34Z","lastTransitionTime":"2025-12-05T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.428784 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.428836 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.428853 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.428874 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.428890 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:34Z","lastTransitionTime":"2025-12-05T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.532242 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.532289 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.532305 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.532326 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.532342 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:34Z","lastTransitionTime":"2025-12-05T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.635611 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.635665 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.635680 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.635701 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.635714 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:34Z","lastTransitionTime":"2025-12-05T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.739170 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.739220 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.739235 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.739255 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.739268 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:34Z","lastTransitionTime":"2025-12-05T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.843054 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.843095 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.843106 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.843122 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.843134 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:34Z","lastTransitionTime":"2025-12-05T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.946074 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.946125 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.946134 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.946152 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:34 crc kubenswrapper[4815]: I1205 09:07:34.946164 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:34Z","lastTransitionTime":"2025-12-05T09:07:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.049642 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.049708 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.049728 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.049752 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.049771 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.152245 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.152302 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.152313 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.152333 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.152346 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.225192 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.225329 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.225429 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.225462 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.225552 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: E1205 09:07:35.238803 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:35Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.241981 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.242008 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.242020 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.242036 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.242047 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: E1205 09:07:35.254542 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:35Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.257901 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.257933 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.257944 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.257961 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.257975 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: E1205 09:07:35.272703 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:35Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.276830 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.276890 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.276904 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.276929 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.276943 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: E1205 09:07:35.292415 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:35Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.298164 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.298355 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.298455 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.298582 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.298673 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: E1205 09:07:35.316159 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:35Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:35 crc kubenswrapper[4815]: E1205 09:07:35.316639 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.318847 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.318900 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.318914 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.318937 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.318951 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.418019 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:35 crc kubenswrapper[4815]: E1205 09:07:35.418149 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.418266 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.418019 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:35 crc kubenswrapper[4815]: E1205 09:07:35.418413 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.418569 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:35 crc kubenswrapper[4815]: E1205 09:07:35.418747 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:35 crc kubenswrapper[4815]: E1205 09:07:35.418736 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.425100 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.425135 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.425146 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.425161 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.425171 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.528214 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.528258 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.528272 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.528287 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.528296 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.630798 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.630860 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.630877 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.630901 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.630923 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.733827 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.733898 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.733921 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.733950 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.733973 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.836234 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.836282 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.836295 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.836311 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.836320 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.938537 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.938575 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.938584 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.938598 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:35 crc kubenswrapper[4815]: I1205 09:07:35.938606 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:35Z","lastTransitionTime":"2025-12-05T09:07:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.040539 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.040575 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.040583 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.040597 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.040605 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:36Z","lastTransitionTime":"2025-12-05T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.143547 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.143612 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.143623 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.143639 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.143652 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:36Z","lastTransitionTime":"2025-12-05T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.246127 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.246172 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.246183 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.246198 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.246209 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:36Z","lastTransitionTime":"2025-12-05T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.348731 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.348790 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.348805 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.348826 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.348843 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:36Z","lastTransitionTime":"2025-12-05T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.451235 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.451332 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.451344 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.451362 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.451373 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:36Z","lastTransitionTime":"2025-12-05T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.554834 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.554889 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.554912 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.554932 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.554945 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:36Z","lastTransitionTime":"2025-12-05T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.657473 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.657554 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.657595 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.657619 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.657631 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:36Z","lastTransitionTime":"2025-12-05T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.760298 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.760583 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.760694 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.760803 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.760899 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:36Z","lastTransitionTime":"2025-12-05T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.864238 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.864304 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.864326 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.864353 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.864375 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:36Z","lastTransitionTime":"2025-12-05T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.967053 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.967102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.967117 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.967134 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:36 crc kubenswrapper[4815]: I1205 09:07:36.967153 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:36Z","lastTransitionTime":"2025-12-05T09:07:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.069697 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.069734 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.069743 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.069755 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.069764 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:37Z","lastTransitionTime":"2025-12-05T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.172085 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.172120 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.172131 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.172148 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.172159 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:37Z","lastTransitionTime":"2025-12-05T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.274105 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.274171 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.274193 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.274221 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.274238 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:37Z","lastTransitionTime":"2025-12-05T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.376347 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.376750 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.376930 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.377091 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.377276 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:37Z","lastTransitionTime":"2025-12-05T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.418611 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.418661 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.419101 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.419790 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:37 crc kubenswrapper[4815]: E1205 09:07:37.419988 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:37 crc kubenswrapper[4815]: E1205 09:07:37.420352 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:37 crc kubenswrapper[4815]: E1205 09:07:37.420466 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:37 crc kubenswrapper[4815]: E1205 09:07:37.420550 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.421337 4815 scope.go:117] "RemoveContainer" containerID="3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.481232 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.481278 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.481294 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.481314 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.481332 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:37Z","lastTransitionTime":"2025-12-05T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.584924 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.584972 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.584986 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.585008 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.585026 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:37Z","lastTransitionTime":"2025-12-05T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.687407 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.687460 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.687470 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.687484 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.687515 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:37Z","lastTransitionTime":"2025-12-05T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.790423 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.790456 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.790467 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.790487 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.790533 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:37Z","lastTransitionTime":"2025-12-05T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.879095 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/1.log" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.882669 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3"} Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.883453 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.892885 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.892921 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.892932 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.892947 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.892957 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:37Z","lastTransitionTime":"2025-12-05T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.898455 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:37Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.910821 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:37Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.923299 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:37Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.939885 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:13Z\\\",\\\"message\\\":\\\"Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:07:12.723214 6171 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 09:07:12.723235 6171 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1205 09:07:12.723240 6171 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:37Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.949824 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:37Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.959334 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:37Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.971828 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:37Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.993597 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:37Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.994998 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.995038 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.995047 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.995062 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:37 crc kubenswrapper[4815]: I1205 09:07:37.995073 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:37Z","lastTransitionTime":"2025-12-05T09:07:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.008163 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.045260 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.063069 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.084236 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.093719 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.097351 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.097389 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.097397 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.097411 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.097421 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:38Z","lastTransitionTime":"2025-12-05T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.108205 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.121032 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.132148 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.145686 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.199528 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.199568 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.199579 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.199595 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.199604 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:38Z","lastTransitionTime":"2025-12-05T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.301553 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.301587 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.301598 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.301611 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.301621 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:38Z","lastTransitionTime":"2025-12-05T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.403838 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.403875 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.403885 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.403898 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.403908 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:38Z","lastTransitionTime":"2025-12-05T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.507332 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.507405 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.507429 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.507460 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.507483 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:38Z","lastTransitionTime":"2025-12-05T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.610026 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.610133 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.610165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.610207 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.610230 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:38Z","lastTransitionTime":"2025-12-05T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.712928 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.712997 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.713020 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.713051 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.713075 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:38Z","lastTransitionTime":"2025-12-05T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.817021 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.817083 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.817099 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.817128 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.817167 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:38Z","lastTransitionTime":"2025-12-05T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.890325 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/2.log" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.891483 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/1.log" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.896539 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3" exitCode=1 Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.896592 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3"} Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.896657 4815 scope.go:117] "RemoveContainer" containerID="3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.897908 4815 scope.go:117] "RemoveContainer" containerID="9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3" Dec 05 09:07:38 crc kubenswrapper[4815]: E1205 09:07:38.898386 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.917796 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.919661 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.919686 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.919697 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.919712 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.919721 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:38Z","lastTransitionTime":"2025-12-05T09:07:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.937562 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.955599 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.972614 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:38 crc kubenswrapper[4815]: I1205 09:07:38.985181 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.012081 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3518b50167b1aadafa364fe4ac2568b34f71c24be9654c7ce8a910ec91ca3d51\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:13Z\\\",\\\"message\\\":\\\"Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:07:12.723214 6171 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]} options:{GoMap:map[iface-id-ver:9d751cbb-f2e2-430d-9754-c882a5e924a5 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:3b 10.217.0.59]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {960d98b2-dc64-4e93-a4b6-9b19847af71e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1205 09:07:12.723235 6171 services_controller.go:452] Built service openshift-kube-apiserver-operator/metrics per-node LB for network=default: []services.LB{}\\\\nF1205 09:07:12.723240 6171 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:11Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:38Z\\\",\\\"message\\\":\\\"ift-network-console/networking-console-plugin-85b44fc459-gdk6g]\\\\nI1205 09:07:38.282770 6444 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1205 09:07:38.282774 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z]\\\\nI1205 09:07:38.282728 6444 services_controller.go:434] Service default/kubernetes retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{kubernetes default 1fcaffea-cfe2-4295-9c2a-a3b3626fb3f1 259 0 2025-02-23 05:11:12 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[component:apiserver provider:kubernetes] map[] [] [] []},Spec:Servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.025233 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.025283 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.025296 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.025316 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.025329 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:39Z","lastTransitionTime":"2025-12-05T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.025613 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.049584 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.067594 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.077855 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.092483 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.110202 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.124687 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.128916 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.128965 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.128978 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.128999 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.129014 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:39Z","lastTransitionTime":"2025-12-05T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.141798 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.160910 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.177965 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.192125 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.231145 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.231194 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.231206 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.231221 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.231232 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:39Z","lastTransitionTime":"2025-12-05T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.334137 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.334179 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.334188 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.334202 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.334212 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:39Z","lastTransitionTime":"2025-12-05T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.418550 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.418585 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.418561 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:39 crc kubenswrapper[4815]: E1205 09:07:39.418693 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:39 crc kubenswrapper[4815]: E1205 09:07:39.418782 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.418806 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:39 crc kubenswrapper[4815]: E1205 09:07:39.418862 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:39 crc kubenswrapper[4815]: E1205 09:07:39.419019 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.437118 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.437180 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.437204 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.437233 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.437252 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:39Z","lastTransitionTime":"2025-12-05T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.540165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.540220 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.540236 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.540258 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.540274 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:39Z","lastTransitionTime":"2025-12-05T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.643117 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.643158 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.643168 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.643181 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.643190 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:39Z","lastTransitionTime":"2025-12-05T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.745526 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.745574 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.745590 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.745611 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.745627 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:39Z","lastTransitionTime":"2025-12-05T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.847931 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.847976 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.847989 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.848006 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.848017 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:39Z","lastTransitionTime":"2025-12-05T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.902666 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/2.log" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.906609 4815 scope.go:117] "RemoveContainer" containerID="9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3" Dec 05 09:07:39 crc kubenswrapper[4815]: E1205 09:07:39.906904 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.924874 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.945114 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.949987 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.950060 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.950085 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.950118 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.950143 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:39Z","lastTransitionTime":"2025-12-05T09:07:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.966566 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:39 crc kubenswrapper[4815]: I1205 09:07:39.986586 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:39Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.005836 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.023144 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.036814 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.048602 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.051968 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.052004 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.052015 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.052032 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.052044 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:40Z","lastTransitionTime":"2025-12-05T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.062421 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.073535 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.085191 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.100070 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.118541 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:38Z\\\",\\\"message\\\":\\\"ift-network-console/networking-console-plugin-85b44fc459-gdk6g]\\\\nI1205 09:07:38.282770 6444 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1205 09:07:38.282774 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z]\\\\nI1205 09:07:38.282728 6444 services_controller.go:434] Service default/kubernetes retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{kubernetes default 1fcaffea-cfe2-4295-9c2a-a3b3626fb3f1 259 0 2025-02-23 05:11:12 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[component:apiserver provider:kubernetes] map[] [] [] []},Spec:Servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.130367 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.148854 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.153978 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.154013 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.154022 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.154037 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.154046 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:40Z","lastTransitionTime":"2025-12-05T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.163178 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.173255 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:40Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.256594 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.256972 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.257154 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.257334 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.257589 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:40Z","lastTransitionTime":"2025-12-05T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.360244 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.360506 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.360596 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.360669 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.360744 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:40Z","lastTransitionTime":"2025-12-05T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.464001 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.464063 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.464081 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.464105 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.464124 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:40Z","lastTransitionTime":"2025-12-05T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.566901 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.567001 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.567024 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.567048 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.567064 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:40Z","lastTransitionTime":"2025-12-05T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.669518 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.669561 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.669576 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.669592 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.669603 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:40Z","lastTransitionTime":"2025-12-05T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.772393 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.772476 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.772542 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.772575 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.772601 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:40Z","lastTransitionTime":"2025-12-05T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.875152 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.875198 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.875210 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.875225 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.875237 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:40Z","lastTransitionTime":"2025-12-05T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.977870 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.977930 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.977939 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.977951 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:40 crc kubenswrapper[4815]: I1205 09:07:40.977962 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:40Z","lastTransitionTime":"2025-12-05T09:07:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.080226 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.080261 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.080269 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.080282 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.080291 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:41Z","lastTransitionTime":"2025-12-05T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.182640 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.182697 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.182709 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.182728 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.182741 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:41Z","lastTransitionTime":"2025-12-05T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.284918 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.284983 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.284994 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.285016 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.285030 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:41Z","lastTransitionTime":"2025-12-05T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.387445 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.387476 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.387486 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.387518 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.387530 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:41Z","lastTransitionTime":"2025-12-05T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.417827 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.417901 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.417896 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.417844 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:41 crc kubenswrapper[4815]: E1205 09:07:41.417980 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:41 crc kubenswrapper[4815]: E1205 09:07:41.418041 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:41 crc kubenswrapper[4815]: E1205 09:07:41.418107 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:41 crc kubenswrapper[4815]: E1205 09:07:41.418163 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.431222 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.449046 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.464405 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.479337 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.490228 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.490538 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.490688 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.490840 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.490967 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:41Z","lastTransitionTime":"2025-12-05T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.493783 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.504687 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.522732 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.539585 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.553278 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.564648 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.579234 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.589818 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.593462 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.593514 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.593525 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.593540 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.593551 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:41Z","lastTransitionTime":"2025-12-05T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.605026 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:38Z\\\",\\\"message\\\":\\\"ift-network-console/networking-console-plugin-85b44fc459-gdk6g]\\\\nI1205 09:07:38.282770 6444 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1205 09:07:38.282774 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z]\\\\nI1205 09:07:38.282728 6444 services_controller.go:434] Service default/kubernetes retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{kubernetes default 1fcaffea-cfe2-4295-9c2a-a3b3626fb3f1 259 0 2025-02-23 05:11:12 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[component:apiserver provider:kubernetes] map[] [] [] []},Spec:Servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.614835 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.623575 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.634417 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.645292 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:41Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.696446 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.696503 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.696511 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.696529 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.696538 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:41Z","lastTransitionTime":"2025-12-05T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.799750 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.799787 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.799800 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.799816 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.799826 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:41Z","lastTransitionTime":"2025-12-05T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.903370 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.903769 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.904252 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.904710 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:41 crc kubenswrapper[4815]: I1205 09:07:41.905078 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:41Z","lastTransitionTime":"2025-12-05T09:07:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.007831 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.008754 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.008932 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.009156 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.009346 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:42Z","lastTransitionTime":"2025-12-05T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.112136 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.112167 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.112178 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.112193 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.112204 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:42Z","lastTransitionTime":"2025-12-05T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.214269 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.214315 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.214326 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.214345 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.214356 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:42Z","lastTransitionTime":"2025-12-05T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.316408 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.316455 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.316468 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.316508 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.316522 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:42Z","lastTransitionTime":"2025-12-05T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.418948 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.418982 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.418991 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.419004 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.419016 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:42Z","lastTransitionTime":"2025-12-05T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.520934 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.520985 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.520997 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.521017 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.521030 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:42Z","lastTransitionTime":"2025-12-05T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.624097 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.624634 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.624846 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.625223 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.625409 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:42Z","lastTransitionTime":"2025-12-05T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.729444 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.730093 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.730243 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.730390 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.730578 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:42Z","lastTransitionTime":"2025-12-05T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.834014 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.834629 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.835714 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.835947 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.836162 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:42Z","lastTransitionTime":"2025-12-05T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.939156 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.939223 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.939246 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.939276 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:42 crc kubenswrapper[4815]: I1205 09:07:42.939302 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:42Z","lastTransitionTime":"2025-12-05T09:07:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.042227 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.042267 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.042275 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.042289 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.042298 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:43Z","lastTransitionTime":"2025-12-05T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.144345 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.144402 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.144418 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.144441 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.144460 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:43Z","lastTransitionTime":"2025-12-05T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.247685 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.247783 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.247806 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.247834 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.247856 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:43Z","lastTransitionTime":"2025-12-05T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.350081 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.350120 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.350140 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.350157 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.350198 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:43Z","lastTransitionTime":"2025-12-05T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.418125 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.418182 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.418134 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:43 crc kubenswrapper[4815]: E1205 09:07:43.418244 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:43 crc kubenswrapper[4815]: E1205 09:07:43.418353 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:43 crc kubenswrapper[4815]: E1205 09:07:43.418427 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.418522 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:43 crc kubenswrapper[4815]: E1205 09:07:43.418581 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.453029 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.453109 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.453130 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.453163 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.453189 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:43Z","lastTransitionTime":"2025-12-05T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.556102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.556144 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.556158 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.556177 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.556190 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:43Z","lastTransitionTime":"2025-12-05T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.657866 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.657894 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.657902 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.657914 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.657922 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:43Z","lastTransitionTime":"2025-12-05T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.759978 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.760008 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.760016 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.760028 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.760035 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:43Z","lastTransitionTime":"2025-12-05T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.863137 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.863197 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.863218 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.863245 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.863264 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:43Z","lastTransitionTime":"2025-12-05T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.965587 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.965613 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.965621 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.965634 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:43 crc kubenswrapper[4815]: I1205 09:07:43.965642 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:43Z","lastTransitionTime":"2025-12-05T09:07:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.067939 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.067975 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.067987 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.068007 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.068019 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:44Z","lastTransitionTime":"2025-12-05T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.170417 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.170445 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.170458 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.170475 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.170506 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:44Z","lastTransitionTime":"2025-12-05T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.272545 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.272574 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.272586 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.272602 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.272612 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:44Z","lastTransitionTime":"2025-12-05T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.375291 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.375334 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.375346 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.375364 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.375375 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:44Z","lastTransitionTime":"2025-12-05T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.477393 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.477469 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.477516 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.477540 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.477559 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:44Z","lastTransitionTime":"2025-12-05T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.580030 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.580070 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.580078 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.580091 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.580401 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:44Z","lastTransitionTime":"2025-12-05T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.682571 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.682614 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.682626 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.682677 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.682692 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:44Z","lastTransitionTime":"2025-12-05T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.785192 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.785246 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.785260 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.785280 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.785295 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:44Z","lastTransitionTime":"2025-12-05T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.888020 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.888065 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.888076 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.888093 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.888107 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:44Z","lastTransitionTime":"2025-12-05T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.990212 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.990249 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.990261 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.990277 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:44 crc kubenswrapper[4815]: I1205 09:07:44.990288 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:44Z","lastTransitionTime":"2025-12-05T09:07:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.092698 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.092741 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.092752 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.092767 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.092777 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.194793 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.194836 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.194844 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.194857 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.194865 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.297068 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.297103 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.297116 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.297133 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.297143 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.380524 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.380565 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.380577 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.380595 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.380604 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: E1205 09:07:45.393383 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.396535 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.396566 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.396574 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.396587 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.396596 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: E1205 09:07:45.406947 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.410770 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.410800 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.410812 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.410824 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.410832 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.419703 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:45 crc kubenswrapper[4815]: E1205 09:07:45.420022 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.419907 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:45 crc kubenswrapper[4815]: E1205 09:07:45.420437 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.419876 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:45 crc kubenswrapper[4815]: E1205 09:07:45.420693 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.419928 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:45 crc kubenswrapper[4815]: E1205 09:07:45.420964 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:45 crc kubenswrapper[4815]: E1205 09:07:45.422764 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.426167 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.426304 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.426384 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.426471 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.426737 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: E1205 09:07:45.438179 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.441062 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.441089 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.441098 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.441115 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.441126 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: E1205 09:07:45.453249 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:45Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:45 crc kubenswrapper[4815]: E1205 09:07:45.453367 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.454720 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.454768 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.454794 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.454810 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.454821 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.557517 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.557554 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.557565 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.557580 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.557592 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.660453 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.660513 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.660569 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.660597 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.660609 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.763296 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.763333 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.763343 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.763358 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.763366 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.866334 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.866372 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.866383 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.866401 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.866414 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.968327 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.968366 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.968377 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.968394 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:45 crc kubenswrapper[4815]: I1205 09:07:45.968405 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:45Z","lastTransitionTime":"2025-12-05T09:07:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.071294 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.071332 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.071346 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.071362 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.071374 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:46Z","lastTransitionTime":"2025-12-05T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.176535 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.176818 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.176930 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.177033 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.177145 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:46Z","lastTransitionTime":"2025-12-05T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.279713 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.279974 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.280056 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.280143 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.280213 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:46Z","lastTransitionTime":"2025-12-05T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.382228 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.382271 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.382283 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.382298 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.382655 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:46Z","lastTransitionTime":"2025-12-05T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.484797 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.484823 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.484831 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.484843 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.484863 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:46Z","lastTransitionTime":"2025-12-05T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.587655 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.587702 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.587715 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.587734 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.587746 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:46Z","lastTransitionTime":"2025-12-05T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.689540 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.689581 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.689593 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.689608 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.689619 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:46Z","lastTransitionTime":"2025-12-05T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.791875 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.791926 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.791938 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.791954 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.791964 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:46Z","lastTransitionTime":"2025-12-05T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.894331 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.894426 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.894445 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.894468 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.894484 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:46Z","lastTransitionTime":"2025-12-05T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.996942 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.996977 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.996989 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.997004 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:46 crc kubenswrapper[4815]: I1205 09:07:46.997015 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:46Z","lastTransitionTime":"2025-12-05T09:07:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.099370 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.099410 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.099423 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.099438 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.099447 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:47Z","lastTransitionTime":"2025-12-05T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.201901 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.201933 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.201945 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.201961 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.201972 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:47Z","lastTransitionTime":"2025-12-05T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.251961 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:47 crc kubenswrapper[4815]: E1205 09:07:47.252141 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:47 crc kubenswrapper[4815]: E1205 09:07:47.252210 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs podName:010248a6-d07e-4e6b-97c8-bda1925d6adb nodeName:}" failed. No retries permitted until 2025-12-05 09:08:19.252187305 +0000 UTC m=+98.130794182 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs") pod "network-metrics-daemon-sslm2" (UID: "010248a6-d07e-4e6b-97c8-bda1925d6adb") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.306119 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.306177 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.306200 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.306228 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.306248 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:47Z","lastTransitionTime":"2025-12-05T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.409025 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.409291 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.409379 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.409459 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.409552 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:47Z","lastTransitionTime":"2025-12-05T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.418332 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.418381 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.418386 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.418603 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:47 crc kubenswrapper[4815]: E1205 09:07:47.418713 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:47 crc kubenswrapper[4815]: E1205 09:07:47.418866 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:47 crc kubenswrapper[4815]: E1205 09:07:47.418933 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:47 crc kubenswrapper[4815]: E1205 09:07:47.419004 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.512309 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.512657 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.512786 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.512926 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.513026 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:47Z","lastTransitionTime":"2025-12-05T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.615779 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.615840 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.615857 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.615878 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.615893 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:47Z","lastTransitionTime":"2025-12-05T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.718908 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.719188 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.719417 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.719682 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.719929 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:47Z","lastTransitionTime":"2025-12-05T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.823838 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.823881 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.823893 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.823912 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.823924 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:47Z","lastTransitionTime":"2025-12-05T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.926506 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.926552 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.926561 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.926593 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:47 crc kubenswrapper[4815]: I1205 09:07:47.926610 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:47Z","lastTransitionTime":"2025-12-05T09:07:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.028925 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.028965 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.028977 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.028994 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.029007 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:48Z","lastTransitionTime":"2025-12-05T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.131559 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.131601 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.131612 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.131629 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.131645 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:48Z","lastTransitionTime":"2025-12-05T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.233512 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.233558 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.233571 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.233589 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.233601 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:48Z","lastTransitionTime":"2025-12-05T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.336403 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.336445 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.336454 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.336468 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.336477 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:48Z","lastTransitionTime":"2025-12-05T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.439649 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.439762 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.439825 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.439854 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.439873 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:48Z","lastTransitionTime":"2025-12-05T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.543250 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.543334 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.543416 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.543438 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.543544 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:48Z","lastTransitionTime":"2025-12-05T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.645987 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.646058 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.646072 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.646089 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.646103 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:48Z","lastTransitionTime":"2025-12-05T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.747707 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.747734 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.747743 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.747755 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.747764 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:48Z","lastTransitionTime":"2025-12-05T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.850206 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.850250 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.850261 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.850277 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.850289 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:48Z","lastTransitionTime":"2025-12-05T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.931847 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-f9lkk_dff16910-bb36-4748-82bb-0f1d90da05b5/kube-multus/0.log" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.931898 4815 generic.go:334] "Generic (PLEG): container finished" podID="dff16910-bb36-4748-82bb-0f1d90da05b5" containerID="f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded" exitCode=1 Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.931929 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-f9lkk" event={"ID":"dff16910-bb36-4748-82bb-0f1d90da05b5","Type":"ContainerDied","Data":"f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.932294 4815 scope.go:117] "RemoveContainer" containerID="f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.952514 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.952540 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.952548 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.952559 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.952567 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:48Z","lastTransitionTime":"2025-12-05T09:07:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.960433 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:48Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.977072 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:48Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.986582 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:48Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:48 crc kubenswrapper[4815]: I1205 09:07:48.995998 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:48Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.006681 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.018356 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.032555 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.046599 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.058580 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.058612 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.058624 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.058638 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.058647 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:49Z","lastTransitionTime":"2025-12-05T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.062380 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.073138 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.097715 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:38Z\\\",\\\"message\\\":\\\"ift-network-console/networking-console-plugin-85b44fc459-gdk6g]\\\\nI1205 09:07:38.282770 6444 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1205 09:07:38.282774 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z]\\\\nI1205 09:07:38.282728 6444 services_controller.go:434] Service default/kubernetes retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{kubernetes default 1fcaffea-cfe2-4295-9c2a-a3b3626fb3f1 259 0 2025-02-23 05:11:12 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[component:apiserver provider:kubernetes] map[] [] [] []},Spec:Servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.110712 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.119801 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.131041 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.142210 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.153748 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.160535 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.160572 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.160583 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.160599 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.160611 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:49Z","lastTransitionTime":"2025-12-05T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.165767 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:48Z\\\",\\\"message\\\":\\\"2025-12-05T09:07:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f\\\\n2025-12-05T09:07:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f to /host/opt/cni/bin/\\\\n2025-12-05T09:07:03Z [verbose] multus-daemon started\\\\n2025-12-05T09:07:03Z [verbose] Readiness Indicator file check\\\\n2025-12-05T09:07:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.263139 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.263170 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.263180 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.263195 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.263204 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:49Z","lastTransitionTime":"2025-12-05T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.365449 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.365483 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.365512 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.365527 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.365536 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:49Z","lastTransitionTime":"2025-12-05T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.418223 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.418294 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:49 crc kubenswrapper[4815]: E1205 09:07:49.418321 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.418339 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.418355 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:49 crc kubenswrapper[4815]: E1205 09:07:49.418473 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:49 crc kubenswrapper[4815]: E1205 09:07:49.418563 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:49 crc kubenswrapper[4815]: E1205 09:07:49.418619 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.467596 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.467637 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.467648 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.467665 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.467677 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:49Z","lastTransitionTime":"2025-12-05T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.569317 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.569343 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.569351 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.569362 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.569371 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:49Z","lastTransitionTime":"2025-12-05T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.672331 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.672362 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.672370 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.672382 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.672392 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:49Z","lastTransitionTime":"2025-12-05T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.775007 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.775067 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.775086 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.775109 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.775126 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:49Z","lastTransitionTime":"2025-12-05T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.877010 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.877053 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.877064 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.877082 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.877095 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:49Z","lastTransitionTime":"2025-12-05T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.936051 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-f9lkk_dff16910-bb36-4748-82bb-0f1d90da05b5/kube-multus/0.log" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.936097 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-f9lkk" event={"ID":"dff16910-bb36-4748-82bb-0f1d90da05b5","Type":"ContainerStarted","Data":"93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b"} Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.955200 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.969142 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.977084 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.978739 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.978767 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.978774 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.978786 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.978795 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:49Z","lastTransitionTime":"2025-12-05T09:07:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.985761 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:49 crc kubenswrapper[4815]: I1205 09:07:49.994614 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:49Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.003665 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.015207 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.029360 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.040080 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.050318 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.065684 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:38Z\\\",\\\"message\\\":\\\"ift-network-console/networking-console-plugin-85b44fc459-gdk6g]\\\\nI1205 09:07:38.282770 6444 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1205 09:07:38.282774 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z]\\\\nI1205 09:07:38.282728 6444 services_controller.go:434] Service default/kubernetes retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{kubernetes default 1fcaffea-cfe2-4295-9c2a-a3b3626fb3f1 259 0 2025-02-23 05:11:12 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[component:apiserver provider:kubernetes] map[] [] [] []},Spec:Servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.076296 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.080228 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.080275 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.080284 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.080302 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.080311 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:50Z","lastTransitionTime":"2025-12-05T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.085906 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.098407 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.108948 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.118344 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.128031 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:48Z\\\",\\\"message\\\":\\\"2025-12-05T09:07:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f\\\\n2025-12-05T09:07:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f to /host/opt/cni/bin/\\\\n2025-12-05T09:07:03Z [verbose] multus-daemon started\\\\n2025-12-05T09:07:03Z [verbose] Readiness Indicator file check\\\\n2025-12-05T09:07:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:50Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.182252 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.182287 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.182298 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.182313 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.182325 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:50Z","lastTransitionTime":"2025-12-05T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.285461 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.285511 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.285519 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.285532 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.285541 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:50Z","lastTransitionTime":"2025-12-05T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.387350 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.387389 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.387405 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.387428 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.387445 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:50Z","lastTransitionTime":"2025-12-05T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.489644 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.489689 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.489707 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.489725 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.489737 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:50Z","lastTransitionTime":"2025-12-05T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.592340 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.592405 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.592427 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.592454 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.592475 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:50Z","lastTransitionTime":"2025-12-05T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.695561 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.695588 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.695598 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.695613 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.695625 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:50Z","lastTransitionTime":"2025-12-05T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.797866 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.797910 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.797923 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.797939 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.797952 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:50Z","lastTransitionTime":"2025-12-05T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.900612 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.900650 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.900659 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.900672 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:50 crc kubenswrapper[4815]: I1205 09:07:50.900682 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:50Z","lastTransitionTime":"2025-12-05T09:07:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.003571 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.003608 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.003617 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.003632 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.003642 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:51Z","lastTransitionTime":"2025-12-05T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.105235 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.105299 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.105316 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.105342 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.105362 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:51Z","lastTransitionTime":"2025-12-05T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.207695 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.207733 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.207745 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.207764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.207781 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:51Z","lastTransitionTime":"2025-12-05T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.310104 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.310139 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.310148 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.310162 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.310171 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:51Z","lastTransitionTime":"2025-12-05T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.412683 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.412742 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.412764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.412796 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.412813 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:51Z","lastTransitionTime":"2025-12-05T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.418137 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.418194 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.418230 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:51 crc kubenswrapper[4815]: E1205 09:07:51.418277 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:51 crc kubenswrapper[4815]: E1205 09:07:51.418441 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:51 crc kubenswrapper[4815]: E1205 09:07:51.418548 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.418591 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:51 crc kubenswrapper[4815]: E1205 09:07:51.418671 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.432218 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.450312 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.466044 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.478542 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.490383 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.502437 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.511522 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.516426 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.516474 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.516514 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.516540 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.516558 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:51Z","lastTransitionTime":"2025-12-05T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.523146 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.533642 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:48Z\\\",\\\"message\\\":\\\"2025-12-05T09:07:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f\\\\n2025-12-05T09:07:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f to /host/opt/cni/bin/\\\\n2025-12-05T09:07:03Z [verbose] multus-daemon started\\\\n2025-12-05T09:07:03Z [verbose] Readiness Indicator file check\\\\n2025-12-05T09:07:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.553024 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:38Z\\\",\\\"message\\\":\\\"ift-network-console/networking-console-plugin-85b44fc459-gdk6g]\\\\nI1205 09:07:38.282770 6444 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1205 09:07:38.282774 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z]\\\\nI1205 09:07:38.282728 6444 services_controller.go:434] Service default/kubernetes retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{kubernetes default 1fcaffea-cfe2-4295-9c2a-a3b3626fb3f1 259 0 2025-02-23 05:11:12 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[component:apiserver provider:kubernetes] map[] [] [] []},Spec:Servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.564013 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.584941 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.598904 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.608185 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.619102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.619293 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.619307 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.619383 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.619405 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:51Z","lastTransitionTime":"2025-12-05T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.620597 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.634456 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.642294 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:51Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.721838 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.721946 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.721958 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.721980 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.721992 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:51Z","lastTransitionTime":"2025-12-05T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.824669 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.824707 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.824718 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.824732 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.824742 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:51Z","lastTransitionTime":"2025-12-05T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.926765 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.926802 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.926814 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.926828 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:51 crc kubenswrapper[4815]: I1205 09:07:51.926838 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:51Z","lastTransitionTime":"2025-12-05T09:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.029038 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.029071 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.029082 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.029097 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.029109 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:52Z","lastTransitionTime":"2025-12-05T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.131478 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.131529 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.131539 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.131556 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.131567 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:52Z","lastTransitionTime":"2025-12-05T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.233958 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.234000 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.234011 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.234027 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.234071 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:52Z","lastTransitionTime":"2025-12-05T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.336440 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.336474 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.336482 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.336516 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.336527 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:52Z","lastTransitionTime":"2025-12-05T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.419300 4815 scope.go:117] "RemoveContainer" containerID="9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3" Dec 05 09:07:52 crc kubenswrapper[4815]: E1205 09:07:52.419622 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.439605 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.439636 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.439645 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.439657 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.439666 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:52Z","lastTransitionTime":"2025-12-05T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.542865 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.542899 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.542908 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.542920 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.542929 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:52Z","lastTransitionTime":"2025-12-05T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.649973 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.650068 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.650084 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.650102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.650530 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:52Z","lastTransitionTime":"2025-12-05T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.752935 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.753074 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.753086 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.753102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.753113 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:52Z","lastTransitionTime":"2025-12-05T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.855050 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.855081 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.855089 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.855102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.855111 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:52Z","lastTransitionTime":"2025-12-05T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.957561 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.957604 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.957615 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.957628 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:52 crc kubenswrapper[4815]: I1205 09:07:52.957638 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:52Z","lastTransitionTime":"2025-12-05T09:07:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.061051 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.061109 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.061119 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.061132 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.061142 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:53Z","lastTransitionTime":"2025-12-05T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.163085 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.163120 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.163132 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.163150 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.163160 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:53Z","lastTransitionTime":"2025-12-05T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.265482 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.265543 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.265557 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.265594 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.265607 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:53Z","lastTransitionTime":"2025-12-05T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.368373 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.368419 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.368429 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.368447 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.368460 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:53Z","lastTransitionTime":"2025-12-05T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.418752 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.418816 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.418860 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.419013 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:53 crc kubenswrapper[4815]: E1205 09:07:53.419112 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:53 crc kubenswrapper[4815]: E1205 09:07:53.419139 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:53 crc kubenswrapper[4815]: E1205 09:07:53.419354 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:53 crc kubenswrapper[4815]: E1205 09:07:53.419417 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.441820 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.471001 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.471081 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.471095 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.471112 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.471124 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:53Z","lastTransitionTime":"2025-12-05T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.573216 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.573273 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.573286 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.573305 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.573315 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:53Z","lastTransitionTime":"2025-12-05T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.676026 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.676088 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.676104 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.676129 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.676146 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:53Z","lastTransitionTime":"2025-12-05T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.778933 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.779014 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.779036 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.779069 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.779091 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:53Z","lastTransitionTime":"2025-12-05T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.882637 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.882694 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.882706 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.882727 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.882739 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:53Z","lastTransitionTime":"2025-12-05T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.987124 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.987176 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.987187 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.987393 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:53 crc kubenswrapper[4815]: I1205 09:07:53.987402 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:53Z","lastTransitionTime":"2025-12-05T09:07:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.089821 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.089858 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.089866 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.089883 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.089893 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:54Z","lastTransitionTime":"2025-12-05T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.192787 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.192847 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.192865 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.192887 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.192904 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:54Z","lastTransitionTime":"2025-12-05T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.299038 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.299126 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.299161 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.299190 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.299210 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:54Z","lastTransitionTime":"2025-12-05T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.401323 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.401390 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.401415 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.401444 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.401461 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:54Z","lastTransitionTime":"2025-12-05T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.505585 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.505626 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.505636 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.505649 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.505657 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:54Z","lastTransitionTime":"2025-12-05T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.608278 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.608337 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.608346 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.608361 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.608370 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:54Z","lastTransitionTime":"2025-12-05T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.712141 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.712181 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.712206 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.712224 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.712234 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:54Z","lastTransitionTime":"2025-12-05T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.815549 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.815619 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.815630 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.815648 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.815659 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:54Z","lastTransitionTime":"2025-12-05T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.918421 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.918462 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.918472 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.918505 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:54 crc kubenswrapper[4815]: I1205 09:07:54.918516 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:54Z","lastTransitionTime":"2025-12-05T09:07:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.020980 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.021030 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.021041 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.021058 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.021071 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.124086 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.124789 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.124923 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.125029 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.125119 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.228085 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.228136 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.228147 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.228165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.228176 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.331775 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.331873 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.331892 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.331917 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.331936 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.418217 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.418319 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.418331 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.418228 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:55 crc kubenswrapper[4815]: E1205 09:07:55.418448 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:55 crc kubenswrapper[4815]: E1205 09:07:55.418609 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:55 crc kubenswrapper[4815]: E1205 09:07:55.418782 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:55 crc kubenswrapper[4815]: E1205 09:07:55.418907 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.434667 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.434709 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.434720 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.434734 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.434746 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.502554 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.502960 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.503092 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.503236 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.503379 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: E1205 09:07:55.517885 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:55Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.522129 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.522361 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.522529 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.522676 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.522858 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: E1205 09:07:55.536328 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:55Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.540433 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.540592 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.540715 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.540797 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.540882 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: E1205 09:07:55.554686 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:55Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.559296 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.559333 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.559346 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.559360 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.559370 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: E1205 09:07:55.577110 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:55Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.580883 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.580903 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.580911 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.580924 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.580932 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: E1205 09:07:55.598794 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:55Z is after 2025-08-24T17:21:41Z" Dec 05 09:07:55 crc kubenswrapper[4815]: E1205 09:07:55.598905 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.600666 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.600693 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.600705 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.600747 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.600763 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.702838 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.702870 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.702877 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.702891 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.702900 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.805553 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.805596 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.805608 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.805623 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.805635 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.909219 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.909332 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.909360 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.909542 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:55 crc kubenswrapper[4815]: I1205 09:07:55.909574 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:55Z","lastTransitionTime":"2025-12-05T09:07:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.013541 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.013608 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.013631 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.013663 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.013687 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:56Z","lastTransitionTime":"2025-12-05T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.122474 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.122579 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.122605 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.122637 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.122660 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:56Z","lastTransitionTime":"2025-12-05T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.225757 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.225825 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.225840 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.225859 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.225873 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:56Z","lastTransitionTime":"2025-12-05T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.329103 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.329215 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.329241 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.329268 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.329289 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:56Z","lastTransitionTime":"2025-12-05T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.432084 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.432149 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.432168 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.432191 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.432208 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:56Z","lastTransitionTime":"2025-12-05T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.535058 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.535129 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.535144 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.535167 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.535182 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:56Z","lastTransitionTime":"2025-12-05T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.637246 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.637307 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.637315 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.637327 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.637338 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:56Z","lastTransitionTime":"2025-12-05T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.739859 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.739901 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.739914 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.739930 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.739941 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:56Z","lastTransitionTime":"2025-12-05T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.842803 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.842849 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.842869 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.842897 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.842919 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:56Z","lastTransitionTime":"2025-12-05T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.945361 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.945399 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.945409 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.945425 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:56 crc kubenswrapper[4815]: I1205 09:07:56.945436 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:56Z","lastTransitionTime":"2025-12-05T09:07:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.048079 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.048433 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.048629 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.048764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.049168 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:57Z","lastTransitionTime":"2025-12-05T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.153630 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.153660 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.153668 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.153684 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.153696 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:57Z","lastTransitionTime":"2025-12-05T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.256719 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.256764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.256781 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.256802 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.256823 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:57Z","lastTransitionTime":"2025-12-05T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.362368 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.362460 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.362574 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.362607 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.362627 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:57Z","lastTransitionTime":"2025-12-05T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.418880 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:57 crc kubenswrapper[4815]: E1205 09:07:57.419092 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.419466 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:57 crc kubenswrapper[4815]: E1205 09:07:57.419671 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.419762 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:57 crc kubenswrapper[4815]: E1205 09:07:57.419975 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.419999 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:57 crc kubenswrapper[4815]: E1205 09:07:57.420199 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.466173 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.466305 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.466378 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.466417 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.466615 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:57Z","lastTransitionTime":"2025-12-05T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.569020 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.569057 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.569069 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.569087 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.569098 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:57Z","lastTransitionTime":"2025-12-05T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.672459 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.672628 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.672653 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.672703 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.672728 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:57Z","lastTransitionTime":"2025-12-05T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.776077 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.776151 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.776219 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.776249 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.776271 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:57Z","lastTransitionTime":"2025-12-05T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.879629 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.879699 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.879720 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.879746 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.879767 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:57Z","lastTransitionTime":"2025-12-05T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.981913 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.981971 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.981988 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.982008 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:57 crc kubenswrapper[4815]: I1205 09:07:57.982036 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:57Z","lastTransitionTime":"2025-12-05T09:07:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.084228 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.084271 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.084284 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.084301 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.084311 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:58Z","lastTransitionTime":"2025-12-05T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.187628 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.187702 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.187724 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.187751 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.187768 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:58Z","lastTransitionTime":"2025-12-05T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.290325 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.290391 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.290403 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.290418 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.290429 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:58Z","lastTransitionTime":"2025-12-05T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.392998 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.393049 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.393059 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.393077 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.393090 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:58Z","lastTransitionTime":"2025-12-05T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.496965 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.497031 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.497054 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.497082 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.497105 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:58Z","lastTransitionTime":"2025-12-05T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.599782 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.599831 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.599855 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.599878 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.599927 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:58Z","lastTransitionTime":"2025-12-05T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.702517 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.702599 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.702645 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.702669 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.702686 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:58Z","lastTransitionTime":"2025-12-05T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.809971 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.810060 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.810085 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.810119 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.810157 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:58Z","lastTransitionTime":"2025-12-05T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.914060 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.914164 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.914182 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.914205 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:58 crc kubenswrapper[4815]: I1205 09:07:58.914222 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:58Z","lastTransitionTime":"2025-12-05T09:07:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.017043 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.017117 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.017141 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.017171 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.017193 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:59Z","lastTransitionTime":"2025-12-05T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.122684 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.122744 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.122766 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.122793 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.122814 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:59Z","lastTransitionTime":"2025-12-05T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.226017 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.226072 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.226085 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.226107 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.226124 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:59Z","lastTransitionTime":"2025-12-05T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.329026 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.329097 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.329121 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.329150 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.329172 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:59Z","lastTransitionTime":"2025-12-05T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.418774 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.418844 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.418853 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:07:59 crc kubenswrapper[4815]: E1205 09:07:59.418989 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:07:59 crc kubenswrapper[4815]: E1205 09:07:59.419059 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:07:59 crc kubenswrapper[4815]: E1205 09:07:59.419209 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.419317 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:07:59 crc kubenswrapper[4815]: E1205 09:07:59.419405 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.431754 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.431795 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.431807 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.431825 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.431839 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:59Z","lastTransitionTime":"2025-12-05T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.534952 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.535009 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.535023 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.535046 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.535060 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:59Z","lastTransitionTime":"2025-12-05T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.638613 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.638681 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.638702 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.638728 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.638750 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:59Z","lastTransitionTime":"2025-12-05T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.742002 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.742277 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.742373 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.742469 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.742591 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:59Z","lastTransitionTime":"2025-12-05T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.846535 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.846579 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.846587 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.846603 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.846612 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:59Z","lastTransitionTime":"2025-12-05T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.949223 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.949546 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.949661 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.949762 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:07:59 crc kubenswrapper[4815]: I1205 09:07:59.949847 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:07:59Z","lastTransitionTime":"2025-12-05T09:07:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.052555 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.052629 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.052648 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.052685 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.052703 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:00Z","lastTransitionTime":"2025-12-05T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.155517 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.155590 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.155616 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.155649 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.155671 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:00Z","lastTransitionTime":"2025-12-05T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.259251 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.259323 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.259341 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.259361 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.259374 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:00Z","lastTransitionTime":"2025-12-05T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.361753 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.361794 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.361803 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.361818 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.361830 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:00Z","lastTransitionTime":"2025-12-05T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.463722 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.463771 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.463793 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.463820 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.463832 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:00Z","lastTransitionTime":"2025-12-05T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.566049 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.566104 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.566124 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.566148 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.566165 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:00Z","lastTransitionTime":"2025-12-05T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.668699 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.668738 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.668758 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.668777 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.668790 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:00Z","lastTransitionTime":"2025-12-05T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.796741 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.796778 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.796789 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.796813 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.796825 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:00Z","lastTransitionTime":"2025-12-05T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.899759 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.899879 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.899901 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.899934 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:00 crc kubenswrapper[4815]: I1205 09:08:00.899957 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:00Z","lastTransitionTime":"2025-12-05T09:08:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.002521 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.002585 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.002595 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.002612 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.002639 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:01Z","lastTransitionTime":"2025-12-05T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.104646 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.104692 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.104703 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.104721 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.104733 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:01Z","lastTransitionTime":"2025-12-05T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.207352 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.207405 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.207416 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.207431 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.207445 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:01Z","lastTransitionTime":"2025-12-05T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.310868 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.310949 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.310960 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.310980 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.310993 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:01Z","lastTransitionTime":"2025-12-05T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.413255 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.413321 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.413337 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.413357 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.413372 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:01Z","lastTransitionTime":"2025-12-05T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.417804 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:01 crc kubenswrapper[4815]: E1205 09:08:01.417983 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.418549 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:01 crc kubenswrapper[4815]: E1205 09:08:01.418641 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.418838 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:01 crc kubenswrapper[4815]: E1205 09:08:01.418924 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.419121 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:01 crc kubenswrapper[4815]: E1205 09:08:01.419189 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.430980 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.444398 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.456133 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.468685 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.484220 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:48Z\\\",\\\"message\\\":\\\"2025-12-05T09:07:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f\\\\n2025-12-05T09:07:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f to /host/opt/cni/bin/\\\\n2025-12-05T09:07:03Z [verbose] multus-daemon started\\\\n2025-12-05T09:07:03Z [verbose] Readiness Indicator file check\\\\n2025-12-05T09:07:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.500247 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:38Z\\\",\\\"message\\\":\\\"ift-network-console/networking-console-plugin-85b44fc459-gdk6g]\\\\nI1205 09:07:38.282770 6444 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1205 09:07:38.282774 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z]\\\\nI1205 09:07:38.282728 6444 services_controller.go:434] Service default/kubernetes retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{kubernetes default 1fcaffea-cfe2-4295-9c2a-a3b3626fb3f1 259 0 2025-02-23 05:11:12 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[component:apiserver provider:kubernetes] map[] [] [] []},Spec:Servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.511540 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.515922 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.515979 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.515990 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.516025 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.516036 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:01Z","lastTransitionTime":"2025-12-05T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.529830 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.545233 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.555767 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.564450 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c206810-e9ee-4480-8d79-8efcad288bb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3392a2d74d29f75e9794efe23bcb2808d70e32561692fbc759a06eb538bf7ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.574406 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.584710 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.593296 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.604670 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.615512 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.618080 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.618129 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.618140 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.618159 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.618169 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:01Z","lastTransitionTime":"2025-12-05T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.631645 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.643908 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:01Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.719832 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.719887 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.719898 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.719914 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.719924 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:01Z","lastTransitionTime":"2025-12-05T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.823228 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.823299 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.823310 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.823328 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.823339 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:01Z","lastTransitionTime":"2025-12-05T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.925879 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.925937 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.925954 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.925978 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:01 crc kubenswrapper[4815]: I1205 09:08:01.926000 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:01Z","lastTransitionTime":"2025-12-05T09:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.029545 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.029618 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.029631 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.029650 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.029662 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:02Z","lastTransitionTime":"2025-12-05T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.132406 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.132549 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.132581 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.132613 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.132670 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:02Z","lastTransitionTime":"2025-12-05T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.235371 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.235423 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.235440 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.235457 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.235469 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:02Z","lastTransitionTime":"2025-12-05T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.337627 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.337695 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.337711 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.337734 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.337751 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:02Z","lastTransitionTime":"2025-12-05T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.440371 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.440412 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.440423 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.440439 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.440451 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:02Z","lastTransitionTime":"2025-12-05T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.543087 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.543114 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.543122 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.543135 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.543144 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:02Z","lastTransitionTime":"2025-12-05T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.645862 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.645935 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.645953 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.645977 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.645994 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:02Z","lastTransitionTime":"2025-12-05T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.749972 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.750027 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.750036 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.750054 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.750066 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:02Z","lastTransitionTime":"2025-12-05T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.853969 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.854027 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.854044 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.854069 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.854088 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:02Z","lastTransitionTime":"2025-12-05T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.957560 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.957630 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.957704 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.957726 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:02 crc kubenswrapper[4815]: I1205 09:08:02.957738 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:02Z","lastTransitionTime":"2025-12-05T09:08:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.061409 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.061469 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.061523 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.061554 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.061578 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:03Z","lastTransitionTime":"2025-12-05T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.164029 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.164063 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.164073 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.164096 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.164110 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:03Z","lastTransitionTime":"2025-12-05T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.266301 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.266350 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.266360 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.266374 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.266385 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:03Z","lastTransitionTime":"2025-12-05T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.368063 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.368091 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.368099 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.368112 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.368120 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:03Z","lastTransitionTime":"2025-12-05T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.417729 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:03 crc kubenswrapper[4815]: E1205 09:08:03.417869 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.418101 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:03 crc kubenswrapper[4815]: E1205 09:08:03.418183 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.418399 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:03 crc kubenswrapper[4815]: E1205 09:08:03.418532 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.418753 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:03 crc kubenswrapper[4815]: E1205 09:08:03.418846 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.471045 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.471077 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.471094 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.471108 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.471119 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:03Z","lastTransitionTime":"2025-12-05T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.573923 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.573971 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.573985 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.574000 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.574011 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:03Z","lastTransitionTime":"2025-12-05T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.677192 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.677271 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.677295 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.677324 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.677346 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:03Z","lastTransitionTime":"2025-12-05T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.780090 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.780165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.780179 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.780201 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.780217 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:03Z","lastTransitionTime":"2025-12-05T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.883645 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.883706 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.883723 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.883749 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.883770 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:03Z","lastTransitionTime":"2025-12-05T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.987242 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.987301 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.987318 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.987343 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:03 crc kubenswrapper[4815]: I1205 09:08:03.987362 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:03Z","lastTransitionTime":"2025-12-05T09:08:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.091856 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.091918 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.091939 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.091965 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.091985 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:04Z","lastTransitionTime":"2025-12-05T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.195742 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.195803 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.195820 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.195844 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.195863 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:04Z","lastTransitionTime":"2025-12-05T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.299452 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.299549 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.299617 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.299650 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.299670 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:04Z","lastTransitionTime":"2025-12-05T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.403534 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.403582 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.403594 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.403614 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.403629 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:04Z","lastTransitionTime":"2025-12-05T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.507744 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.507843 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.507865 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.507889 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.507947 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:04Z","lastTransitionTime":"2025-12-05T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.611693 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.611769 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.611793 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.611824 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.611846 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:04Z","lastTransitionTime":"2025-12-05T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.714344 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.714380 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.714392 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.714408 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.714418 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:04Z","lastTransitionTime":"2025-12-05T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.816649 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.816703 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.816716 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.816734 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.816746 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:04Z","lastTransitionTime":"2025-12-05T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.920403 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.920458 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.920522 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.920550 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:04 crc kubenswrapper[4815]: I1205 09:08:04.920568 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:04Z","lastTransitionTime":"2025-12-05T09:08:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.023517 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.023864 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.023876 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.023895 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.023908 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.126192 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.126229 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.126240 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.126256 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.126267 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.230009 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.230050 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.230099 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.230119 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.230133 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.332732 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.332773 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.332785 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.332804 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.332817 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.344583 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.344929 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:09.344893231 +0000 UTC m=+148.223500108 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.418660 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.418784 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.418953 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.419251 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.419355 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.419585 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.419747 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.419855 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.436233 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.436300 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.436322 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.436348 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.436371 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.445767 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.445836 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.445926 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.445997 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446074 4815 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446091 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446093 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446127 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:09:09.446110605 +0000 UTC m=+148.324717442 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446125 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446145 4815 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446152 4815 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446157 4815 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446167 4815 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446214 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-05 09:09:09.446196897 +0000 UTC m=+148.324803774 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446254 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-05 09:09:09.446228328 +0000 UTC m=+148.324835205 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.446288 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-05 09:09:09.44627208 +0000 UTC m=+148.324878957 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.539630 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.539701 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.539714 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.539730 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.539743 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.642973 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.643052 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.643085 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.643108 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.643126 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.745697 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.745787 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.745804 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.745828 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.745845 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.848771 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.848856 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.848879 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.848925 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.849031 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.924022 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.924089 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.924102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.924117 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.924129 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.944591 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.949390 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.949447 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.949459 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.949512 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.949523 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.965848 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.969841 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.969892 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.969905 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.969926 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.969939 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:05 crc kubenswrapper[4815]: E1205 09:08:05.985135 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:05Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.989616 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.989671 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.989684 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.989707 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:05 crc kubenswrapper[4815]: I1205 09:08:05.989720 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:05Z","lastTransitionTime":"2025-12-05T09:08:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:06 crc kubenswrapper[4815]: E1205 09:08:06.005280 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.010169 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.010225 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.010238 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.010257 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.010271 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:06Z","lastTransitionTime":"2025-12-05T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:06 crc kubenswrapper[4815]: E1205 09:08:06.026433 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:06Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:06 crc kubenswrapper[4815]: E1205 09:08:06.026637 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.028520 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.028578 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.028597 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.028622 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.028640 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:06Z","lastTransitionTime":"2025-12-05T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.131893 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.131943 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.131964 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.131983 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.131997 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:06Z","lastTransitionTime":"2025-12-05T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.234395 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.234437 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.234446 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.234460 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.234469 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:06Z","lastTransitionTime":"2025-12-05T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.337942 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.337993 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.338006 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.338025 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.338038 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:06Z","lastTransitionTime":"2025-12-05T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.440370 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.440508 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.441689 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.441718 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.441732 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:06Z","lastTransitionTime":"2025-12-05T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.544294 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.544392 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.544414 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.544902 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.545328 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:06Z","lastTransitionTime":"2025-12-05T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.648110 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.648159 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.648173 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.648190 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.648204 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:06Z","lastTransitionTime":"2025-12-05T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.751324 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.751375 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.751394 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.751416 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.751431 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:06Z","lastTransitionTime":"2025-12-05T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.854709 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.854771 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.854788 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.854813 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.854830 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:06Z","lastTransitionTime":"2025-12-05T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.957785 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.957824 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.957836 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.957867 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:06 crc kubenswrapper[4815]: I1205 09:08:06.957879 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:06Z","lastTransitionTime":"2025-12-05T09:08:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.061070 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.061119 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.061130 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.061148 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.061160 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:07Z","lastTransitionTime":"2025-12-05T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.164573 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.164605 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.164617 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.164631 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.164643 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:07Z","lastTransitionTime":"2025-12-05T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.268744 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.268897 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.268925 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.268952 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.268973 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:07Z","lastTransitionTime":"2025-12-05T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.371859 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.371922 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.371946 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.371974 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.371992 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:07Z","lastTransitionTime":"2025-12-05T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.418369 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.418457 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:07 crc kubenswrapper[4815]: E1205 09:08:07.418587 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.418518 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.418611 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:07 crc kubenswrapper[4815]: E1205 09:08:07.418735 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:07 crc kubenswrapper[4815]: E1205 09:08:07.418928 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:07 crc kubenswrapper[4815]: E1205 09:08:07.419294 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.419566 4815 scope.go:117] "RemoveContainer" containerID="9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.474978 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.475047 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.475070 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.475109 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.475130 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:07Z","lastTransitionTime":"2025-12-05T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.578172 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.578217 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.578228 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.578245 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.578258 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:07Z","lastTransitionTime":"2025-12-05T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.680224 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.680265 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.680275 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.680292 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.680303 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:07Z","lastTransitionTime":"2025-12-05T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.781986 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.782026 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.782038 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.782053 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.782065 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:07Z","lastTransitionTime":"2025-12-05T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.884601 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.884646 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.884655 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.884675 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.884686 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:07Z","lastTransitionTime":"2025-12-05T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.986968 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.987013 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.987024 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.987039 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.987049 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:07Z","lastTransitionTime":"2025-12-05T09:08:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.995077 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/2.log" Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.999263 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca"} Dec 05 09:08:07 crc kubenswrapper[4815]: I1205 09:08:07.999721 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.010739 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c206810-e9ee-4480-8d79-8efcad288bb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3392a2d74d29f75e9794efe23bcb2808d70e32561692fbc759a06eb538bf7ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.025474 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.045167 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.057773 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.073428 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.088582 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.089723 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.089750 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.089762 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.089779 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.089791 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:08Z","lastTransitionTime":"2025-12-05T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.102346 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.119386 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.136043 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.148106 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.164074 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.177909 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:48Z\\\",\\\"message\\\":\\\"2025-12-05T09:07:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f\\\\n2025-12-05T09:07:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f to /host/opt/cni/bin/\\\\n2025-12-05T09:07:03Z [verbose] multus-daemon started\\\\n2025-12-05T09:07:03Z [verbose] Readiness Indicator file check\\\\n2025-12-05T09:07:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.191764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.191802 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.191811 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.191825 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.191834 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:08Z","lastTransitionTime":"2025-12-05T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.195448 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:38Z\\\",\\\"message\\\":\\\"ift-network-console/networking-console-plugin-85b44fc459-gdk6g]\\\\nI1205 09:07:38.282770 6444 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1205 09:07:38.282774 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z]\\\\nI1205 09:07:38.282728 6444 services_controller.go:434] Service default/kubernetes retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{kubernetes default 1fcaffea-cfe2-4295-9c2a-a3b3626fb3f1 259 0 2025-02-23 05:11:12 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[component:apiserver provider:kubernetes] map[] [] [] []},Spec:Servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.207286 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.220400 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.239440 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.252472 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.260776 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:08Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.293906 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.293996 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.294016 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.294032 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.294044 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:08Z","lastTransitionTime":"2025-12-05T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.396270 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.396300 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.396312 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.396327 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.396339 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:08Z","lastTransitionTime":"2025-12-05T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.499204 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.499281 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.499305 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.499334 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.499357 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:08Z","lastTransitionTime":"2025-12-05T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.602020 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.602062 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.602073 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.602093 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.602112 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:08Z","lastTransitionTime":"2025-12-05T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.705088 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.705134 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.705147 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.705166 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.705178 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:08Z","lastTransitionTime":"2025-12-05T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.807567 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.807611 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.807625 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.807645 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.807659 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:08Z","lastTransitionTime":"2025-12-05T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.910639 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.910708 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.910725 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.910749 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:08 crc kubenswrapper[4815]: I1205 09:08:08.910766 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:08Z","lastTransitionTime":"2025-12-05T09:08:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.006096 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/3.log" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.007404 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/2.log" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.011285 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca" exitCode=1 Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.011372 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca"} Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.011461 4815 scope.go:117] "RemoveContainer" containerID="9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.012434 4815 scope.go:117] "RemoveContainer" containerID="c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca" Dec 05 09:08:09 crc kubenswrapper[4815]: E1205 09:08:09.012672 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.012870 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.012897 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.012910 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.012931 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.012944 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:09Z","lastTransitionTime":"2025-12-05T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.040896 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.057616 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.069015 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.082982 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c206810-e9ee-4480-8d79-8efcad288bb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3392a2d74d29f75e9794efe23bcb2808d70e32561692fbc759a06eb538bf7ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.098776 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.115116 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.116195 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.116254 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.116267 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.116284 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.116297 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:09Z","lastTransitionTime":"2025-12-05T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.128381 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.145118 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.157954 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.173077 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.187523 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.199878 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.212655 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.218439 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.218528 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.218543 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.218563 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.218600 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:09Z","lastTransitionTime":"2025-12-05T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.224521 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.237164 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.249315 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.260724 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:48Z\\\",\\\"message\\\":\\\"2025-12-05T09:07:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f\\\\n2025-12-05T09:07:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f to /host/opt/cni/bin/\\\\n2025-12-05T09:07:03Z [verbose] multus-daemon started\\\\n2025-12-05T09:07:03Z [verbose] Readiness Indicator file check\\\\n2025-12-05T09:07:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.281187 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9b3c82ee96ac2a98f01bec811ff21faa2747a57ce41b7ad87125578dbc5498c3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:38Z\\\",\\\"message\\\":\\\"ift-network-console/networking-console-plugin-85b44fc459-gdk6g]\\\\nI1205 09:07:38.282770 6444 obj_retry.go:418] Waiting for all the *v1.Pod retry setup to complete in iterateRetryResources\\\\nF1205 09:07:38.282774 6444 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:07:38Z is after 2025-08-24T17:21:41Z]\\\\nI1205 09:07:38.282728 6444 services_controller.go:434] Service default/kubernetes retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{kubernetes default 1fcaffea-cfe2-4295-9c2a-a3b3626fb3f1 259 0 2025-02-23 05:11:12 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[component:apiserver provider:kubernetes] map[] [] [] []},Spec:Servic\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:37Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:08:08Z\\\",\\\"message\\\":\\\"TimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:08:08.220879 6857 services_controller.go:445] Built service default/kubernetes LB template configs for network=default: []services.lbConfig(nil)\\\\nI1205 09:08:08.220881 6857 services_controller.go:452] Built service openshift-service-ca-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1205 09:08:08.220888 6857 services_controller.go:453] Built service openshift-service-ca-operator/metrics template LB for network=default: []services.LB{}\\\\nI1205 09:08:08.220895 6857 services_controller.go:454] Service openshift-service-ca-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1205 09:08:08.220897 6857 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:08:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:09Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.320318 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.320438 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.320525 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.320588 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.320644 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:09Z","lastTransitionTime":"2025-12-05T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.418530 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.418536 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.418625 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:09 crc kubenswrapper[4815]: E1205 09:08:09.418737 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:09 crc kubenswrapper[4815]: E1205 09:08:09.418903 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:09 crc kubenswrapper[4815]: E1205 09:08:09.418980 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.419156 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:09 crc kubenswrapper[4815]: E1205 09:08:09.419321 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.423267 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.423309 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.423320 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.423336 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.423349 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:09Z","lastTransitionTime":"2025-12-05T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.525675 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.525716 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.525728 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.525748 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.525767 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:09Z","lastTransitionTime":"2025-12-05T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.628541 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.628580 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.628588 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.628604 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.628613 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:09Z","lastTransitionTime":"2025-12-05T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.732115 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.732165 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.732179 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.732198 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.732213 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:09Z","lastTransitionTime":"2025-12-05T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.834997 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.835053 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.835069 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.835091 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.835108 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:09Z","lastTransitionTime":"2025-12-05T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.937981 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.938059 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.938076 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.938099 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:09 crc kubenswrapper[4815]: I1205 09:08:09.938116 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:09Z","lastTransitionTime":"2025-12-05T09:08:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.016161 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/3.log" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.021438 4815 scope.go:117] "RemoveContainer" containerID="c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca" Dec 05 09:08:10 crc kubenswrapper[4815]: E1205 09:08:10.021764 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.040678 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.040730 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.040747 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.040767 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.040782 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:10Z","lastTransitionTime":"2025-12-05T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.046836 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.062206 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.073359 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.085019 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.094912 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.105658 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c206810-e9ee-4480-8d79-8efcad288bb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3392a2d74d29f75e9794efe23bcb2808d70e32561692fbc759a06eb538bf7ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.118305 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.127910 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.142326 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.142811 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.142830 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.142838 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.142851 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.142859 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:10Z","lastTransitionTime":"2025-12-05T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.154988 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.170147 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.184306 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:48Z\\\",\\\"message\\\":\\\"2025-12-05T09:07:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f\\\\n2025-12-05T09:07:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f to /host/opt/cni/bin/\\\\n2025-12-05T09:07:03Z [verbose] multus-daemon started\\\\n2025-12-05T09:07:03Z [verbose] Readiness Indicator file check\\\\n2025-12-05T09:07:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.202397 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:08:08Z\\\",\\\"message\\\":\\\"TimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:08:08.220879 6857 services_controller.go:445] Built service default/kubernetes LB template configs for network=default: []services.lbConfig(nil)\\\\nI1205 09:08:08.220881 6857 services_controller.go:452] Built service openshift-service-ca-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1205 09:08:08.220888 6857 services_controller.go:453] Built service openshift-service-ca-operator/metrics template LB for network=default: []services.LB{}\\\\nI1205 09:08:08.220895 6857 services_controller.go:454] Service openshift-service-ca-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1205 09:08:08.220897 6857 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:08:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.217768 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.232160 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.244876 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.244905 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.244913 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.244926 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.244935 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:10Z","lastTransitionTime":"2025-12-05T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.247655 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.262162 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.275277 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:10Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.347839 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.347884 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.347897 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.347918 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.347934 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:10Z","lastTransitionTime":"2025-12-05T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.451745 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.451803 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.451821 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.451846 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.451863 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:10Z","lastTransitionTime":"2025-12-05T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.554643 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.554710 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.554762 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.554790 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.554809 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:10Z","lastTransitionTime":"2025-12-05T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.658175 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.658233 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.658247 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.658262 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.658273 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:10Z","lastTransitionTime":"2025-12-05T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.761577 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.761660 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.761678 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.761702 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.761719 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:10Z","lastTransitionTime":"2025-12-05T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.864403 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.864506 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.864522 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.864539 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.864551 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:10Z","lastTransitionTime":"2025-12-05T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.967848 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.967895 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.967904 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.967920 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:10 crc kubenswrapper[4815]: I1205 09:08:10.967931 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:10Z","lastTransitionTime":"2025-12-05T09:08:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.071383 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.071423 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.071434 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.071451 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.071463 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:11Z","lastTransitionTime":"2025-12-05T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.174883 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.174943 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.174961 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.174984 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.175001 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:11Z","lastTransitionTime":"2025-12-05T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.277333 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.277363 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.277372 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.277384 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.277392 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:11Z","lastTransitionTime":"2025-12-05T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.380406 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.380456 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.380473 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.380526 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.380544 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:11Z","lastTransitionTime":"2025-12-05T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.418211 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.418342 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.418307 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:11 crc kubenswrapper[4815]: E1205 09:08:11.418597 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.418642 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:11 crc kubenswrapper[4815]: E1205 09:08:11.418682 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:11 crc kubenswrapper[4815]: E1205 09:08:11.418836 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:11 crc kubenswrapper[4815]: E1205 09:08:11.418931 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.433384 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.437149 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5e44bbe8-6099-453d-8d03-275929053b58\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1205 09:06:54.960233 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1205 09:06:54.962259 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1562371270/tls.crt::/tmp/serving-cert-1562371270/tls.key\\\\\\\"\\\\nI1205 09:07:00.903856 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1205 09:07:00.906935 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1205 09:07:00.907060 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1205 09:07:00.907110 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1205 09:07:00.907140 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1205 09:07:00.912159 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1205 09:07:00.912196 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912201 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1205 09:07:00.912205 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1205 09:07:00.912207 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1205 09:07:00.912210 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1205 09:07:00.912213 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1205 09:07:00.912434 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1205 09:07:00.913768 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.455444 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.476566 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.483313 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.483349 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.483360 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.483377 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.483389 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:11Z","lastTransitionTime":"2025-12-05T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.488799 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.506029 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.520338 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://109998256bbb340482757dafaaa281fe7560a84839c3f6cad657fdd3772f5235\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.540189 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09908eedcbdc70066ba7fd7d742198d05f8e595643236716da24e920fc91105a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c27846d9bf5cdd1ced7b70e5d23219f5df18267a49fc213fd2a97b8b61165bd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.559467 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-f9lkk" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"dff16910-bb36-4748-82bb-0f1d90da05b5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:49Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:07:48Z\\\",\\\"message\\\":\\\"2025-12-05T09:07:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f\\\\n2025-12-05T09:07:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_66b24171-1211-42a5-90fc-20dda6ea9d8f to /host/opt/cni/bin/\\\\n2025-12-05T09:07:03Z [verbose] multus-daemon started\\\\n2025-12-05T09:07:03Z [verbose] Readiness Indicator file check\\\\n2025-12-05T09:07:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tv78q\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-f9lkk\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.580779 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e930cfad-5770-4db6-826b-353554e2f23c\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-05T09:08:08Z\\\",\\\"message\\\":\\\"TimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.40\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1205 09:08:08.220879 6857 services_controller.go:445] Built service default/kubernetes LB template configs for network=default: []services.lbConfig(nil)\\\\nI1205 09:08:08.220881 6857 services_controller.go:452] Built service openshift-service-ca-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI1205 09:08:08.220888 6857 services_controller.go:453] Built service openshift-service-ca-operator/metrics template LB for network=default: []services.LB{}\\\\nI1205 09:08:08.220895 6857 services_controller.go:454] Service openshift-service-ca-operator/metrics for network=default has 1 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nF1205 09:08:08.220897 6857 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handle\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-05T09:08:07Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bgqrd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-pzlxw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.586012 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.586071 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.586083 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.586099 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.586108 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:11Z","lastTransitionTime":"2025-12-05T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.593127 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd567f9e-eb99-4085-97ee-63a58de73556\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f622e8a0284c3f27d517be6284a0ba3696246ef4b92775bbff2b9a7706458f6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6b0b12b717480f3252ecff32fcc07b101d3cdd9161196e1a8220d29ca0d1a2ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hsxf9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-xd2x5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.601551 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-sslm2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"010248a6-d07e-4e6b-97c8-bda1925d6adb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vmrzd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-sslm2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.620024 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fbcd06bc-cd1d-4f6b-ae48-41232b1412b8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://247a2550ce4fab574dc523dfca8469ba179e61ae46ec7b157c33ebf22483d755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5733552e6951c55b34847824691ab27bb78333f48b0e6a06bbab5c9655bbdaf1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d37edfb74ca7912e5b35d38e1b03bdd55243894020872872b26959e5f20e993\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82cfc97b440ab8d03f971ab601deed98032ddd71c4f5ad62232ab04c7c8be79\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8018c8fbdc55e08ca3654da9ac51c46be09539d653a9e08bccc489452e9d48fd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9e16bca9f88a97fba58bdcc33e59f7a6c1942e73bcf7f351c2d42ff91747cb39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f5247906b6ed399c23e5ecf10c5dd529faffb565657bb9785e81b2b63a8cc94c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://641a4d5ecd13ab5287f1a0a74266eeea62a41bdd891057418bba3fffb60990c7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.636242 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vrs47" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a27419f-8364-4ff9-a6bd-677c3fa72fef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://343bb3e0c4c324f796a6c19780c8a8a0058848b948b97d18e47b0b440aa47237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e7f53685fe01ea20d17b1fc0fe617bb6900cad2a9411f925686eda294eb71292\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9de2d66cce0b3ddf55492575e2feb23f1bbf52fb593810e15c5d2f79aef29c8b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://938a7e699aa35942ef17ab267c02368731d53bdd40e4e468d44a5ed73af91350\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0d1ee965e59035276e8abdae5227251790179abc4df068083e5b5664d94c74b2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5181b6f2791876844ac8c2f659fd96b3314c22659185ae7838a0cf0927c957df\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://70a959dacb792bce697fcf9841724e4be4ec14205fa91b36564fb15388ea4ddd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-clbvz\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vrs47\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.647029 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-625kt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9193a1c8-5c57-4bec-a4cf-3fc1b7e6be57\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://24487027891dbf660aa5a0b91b9e8c55e65217d8baaf38c6ce40cdb260efa4cc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ljd7h\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:06Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-625kt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.659142 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c206810-e9ee-4480-8d79-8efcad288bb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3392a2d74d29f75e9794efe23bcb2808d70e32561692fbc759a06eb538bf7ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.672550 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.686066 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.687561 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.687589 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.687600 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.687615 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.687627 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:11Z","lastTransitionTime":"2025-12-05T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.696268 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:11Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.789705 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.789787 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.789800 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.789815 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.789824 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:11Z","lastTransitionTime":"2025-12-05T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.892877 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.892921 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.892953 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.892971 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.892983 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:11Z","lastTransitionTime":"2025-12-05T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.994589 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.994617 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.994625 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.994639 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:11 crc kubenswrapper[4815]: I1205 09:08:11.994650 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:11Z","lastTransitionTime":"2025-12-05T09:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.097739 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.097777 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.097786 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.097800 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.097810 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:12Z","lastTransitionTime":"2025-12-05T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.200679 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.200731 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.200741 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.200758 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.200769 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:12Z","lastTransitionTime":"2025-12-05T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.303560 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.303623 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.303645 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.303671 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.303692 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:12Z","lastTransitionTime":"2025-12-05T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.405648 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.405716 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.405736 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.405764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.405785 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:12Z","lastTransitionTime":"2025-12-05T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.508214 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.508273 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.508295 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.508323 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.508343 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:12Z","lastTransitionTime":"2025-12-05T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.611516 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.611560 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.611570 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.611585 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.611595 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:12Z","lastTransitionTime":"2025-12-05T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.715373 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.715453 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.715469 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.715522 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.715539 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:12Z","lastTransitionTime":"2025-12-05T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.818894 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.818944 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.818957 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.818975 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.818988 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:12Z","lastTransitionTime":"2025-12-05T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.922746 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.922778 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.922787 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.922800 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:12 crc kubenswrapper[4815]: I1205 09:08:12.922808 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:12Z","lastTransitionTime":"2025-12-05T09:08:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.024957 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.025018 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.025036 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.025060 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.025077 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:13Z","lastTransitionTime":"2025-12-05T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.129048 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.129120 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.129139 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.129163 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.129181 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:13Z","lastTransitionTime":"2025-12-05T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.231816 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.231899 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.232126 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.232157 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.232180 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:13Z","lastTransitionTime":"2025-12-05T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.333905 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.333950 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.333962 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.333980 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.333991 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:13Z","lastTransitionTime":"2025-12-05T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.417833 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:13 crc kubenswrapper[4815]: E1205 09:08:13.418020 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.418106 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.418203 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:13 crc kubenswrapper[4815]: E1205 09:08:13.418306 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.418215 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:13 crc kubenswrapper[4815]: E1205 09:08:13.418384 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:13 crc kubenswrapper[4815]: E1205 09:08:13.418448 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.435908 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.435942 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.435950 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.435963 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.435972 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:13Z","lastTransitionTime":"2025-12-05T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.538321 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.538408 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.538432 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.538462 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.538527 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:13Z","lastTransitionTime":"2025-12-05T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.659095 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.659159 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.659175 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.659201 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.659219 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:13Z","lastTransitionTime":"2025-12-05T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.761611 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.761674 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.761689 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.761707 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.761721 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:13Z","lastTransitionTime":"2025-12-05T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.864591 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.864637 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.864651 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.864668 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.864681 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:13Z","lastTransitionTime":"2025-12-05T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.967362 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.967427 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.967446 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.967470 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:13 crc kubenswrapper[4815]: I1205 09:08:13.967522 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:13Z","lastTransitionTime":"2025-12-05T09:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.069859 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.069896 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.069908 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.069923 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.069940 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:14Z","lastTransitionTime":"2025-12-05T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.172898 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.172934 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.172945 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.172964 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.172976 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:14Z","lastTransitionTime":"2025-12-05T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.275446 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.275536 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.275550 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.275578 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.275592 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:14Z","lastTransitionTime":"2025-12-05T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.378748 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.378802 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.378813 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.378834 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.378846 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:14Z","lastTransitionTime":"2025-12-05T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.481551 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.481615 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.481637 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.481665 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.481686 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:14Z","lastTransitionTime":"2025-12-05T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.584676 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.584740 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.584760 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.584786 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.584801 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:14Z","lastTransitionTime":"2025-12-05T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.687689 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.687755 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.687779 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.687807 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.687828 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:14Z","lastTransitionTime":"2025-12-05T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.791613 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.791692 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.791724 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.791793 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.791813 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:14Z","lastTransitionTime":"2025-12-05T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.895444 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.895558 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.895584 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.895612 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.895633 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:14Z","lastTransitionTime":"2025-12-05T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.998200 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.998274 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.998296 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.998324 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:14 crc kubenswrapper[4815]: I1205 09:08:14.998345 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:14Z","lastTransitionTime":"2025-12-05T09:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.101968 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.102043 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.102065 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.102093 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.102117 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:15Z","lastTransitionTime":"2025-12-05T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.207533 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.207615 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.207640 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.207674 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.207707 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:15Z","lastTransitionTime":"2025-12-05T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.311061 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.311128 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.311153 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.311183 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.311206 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:15Z","lastTransitionTime":"2025-12-05T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.413558 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.413631 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.413652 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.413675 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.413692 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:15Z","lastTransitionTime":"2025-12-05T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.417810 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.417922 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.417839 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:15 crc kubenswrapper[4815]: E1205 09:08:15.418035 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.418082 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:15 crc kubenswrapper[4815]: E1205 09:08:15.418187 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:15 crc kubenswrapper[4815]: E1205 09:08:15.418396 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:15 crc kubenswrapper[4815]: E1205 09:08:15.418545 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.517027 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.517100 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.517118 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.517144 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.517161 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:15Z","lastTransitionTime":"2025-12-05T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.620162 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.620207 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.620217 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.620228 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.620237 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:15Z","lastTransitionTime":"2025-12-05T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.722853 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.722899 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.722909 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.722946 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.722957 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:15Z","lastTransitionTime":"2025-12-05T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.825909 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.825963 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.825978 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.826003 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.826021 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:15Z","lastTransitionTime":"2025-12-05T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.929714 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.929777 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.929800 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.929827 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:15 crc kubenswrapper[4815]: I1205 09:08:15.929847 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:15Z","lastTransitionTime":"2025-12-05T09:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.032947 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.033016 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.033034 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.033088 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.033106 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.136662 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.136738 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.136764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.136795 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.136816 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.239539 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.239572 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.239582 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.239596 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.239607 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.329444 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.329584 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.329619 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.329649 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.329670 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: E1205 09:08:16.345279 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:16Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.349811 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.349848 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.349858 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.349873 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.349884 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: E1205 09:08:16.365323 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:16Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.369620 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.369661 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.369672 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.369688 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.369699 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: E1205 09:08:16.383350 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:16Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.387660 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.387695 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.387706 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.387722 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.387732 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: E1205 09:08:16.404254 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:16Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.409289 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.409334 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.409356 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.409414 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.409427 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: E1205 09:08:16.424303 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-05T09:08:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"e9726a9b-025c-49da-b818-99a967ee21f0\\\",\\\"systemUUID\\\":\\\"25928f9c-29aa-4c55-87c5-643012e42318\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:16Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:16 crc kubenswrapper[4815]: E1205 09:08:16.424530 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.426553 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.426635 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.426653 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.426698 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.426715 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.529618 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.529698 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.529709 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.529724 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.529734 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.633411 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.633458 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.633466 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.633479 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.633530 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.736568 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.736625 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.736645 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.736666 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.736682 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.838731 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.838769 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.838778 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.838791 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.838801 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.940623 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.940674 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.940685 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.940701 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:16 crc kubenswrapper[4815]: I1205 09:08:16.940709 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:16Z","lastTransitionTime":"2025-12-05T09:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.043169 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.043272 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.043293 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.043320 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.043339 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:17Z","lastTransitionTime":"2025-12-05T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.145987 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.146260 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.146345 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.146449 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.146561 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:17Z","lastTransitionTime":"2025-12-05T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.249666 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.249961 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.250082 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.250201 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.250350 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:17Z","lastTransitionTime":"2025-12-05T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.352966 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.353211 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.353284 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.353374 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.353459 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:17Z","lastTransitionTime":"2025-12-05T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.417831 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.417828 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:17 crc kubenswrapper[4815]: E1205 09:08:17.418381 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.417963 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:17 crc kubenswrapper[4815]: E1205 09:08:17.418749 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.417884 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:17 crc kubenswrapper[4815]: E1205 09:08:17.418991 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:17 crc kubenswrapper[4815]: E1205 09:08:17.418477 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.455462 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.455544 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.455553 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.455568 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.455577 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:17Z","lastTransitionTime":"2025-12-05T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.565230 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.565561 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.565699 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.565805 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.565898 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:17Z","lastTransitionTime":"2025-12-05T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.667986 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.668028 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.668038 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.668053 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.668063 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:17Z","lastTransitionTime":"2025-12-05T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.770968 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.771009 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.771028 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.771052 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.771072 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:17Z","lastTransitionTime":"2025-12-05T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.873378 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.873420 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.873451 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.873468 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.873479 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:17Z","lastTransitionTime":"2025-12-05T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.975849 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.975903 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.975912 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.975924 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:17 crc kubenswrapper[4815]: I1205 09:08:17.975932 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:17Z","lastTransitionTime":"2025-12-05T09:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.078852 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.078888 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.078898 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.078916 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.078929 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:18Z","lastTransitionTime":"2025-12-05T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.181382 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.181422 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.181432 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.181451 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.181462 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:18Z","lastTransitionTime":"2025-12-05T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.283870 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.283924 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.283940 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.283964 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.283981 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:18Z","lastTransitionTime":"2025-12-05T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.386203 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.386280 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.386302 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.386333 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.386369 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:18Z","lastTransitionTime":"2025-12-05T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.489785 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.489819 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.489827 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.489840 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.489848 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:18Z","lastTransitionTime":"2025-12-05T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.591859 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.591889 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.591896 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.591911 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.591922 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:18Z","lastTransitionTime":"2025-12-05T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.695208 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.695240 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.695248 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.695261 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.695270 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:18Z","lastTransitionTime":"2025-12-05T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.798207 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.798387 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.798472 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.798561 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.798586 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:18Z","lastTransitionTime":"2025-12-05T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.901070 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.901130 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.901140 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.901160 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:18 crc kubenswrapper[4815]: I1205 09:08:18.901171 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:18Z","lastTransitionTime":"2025-12-05T09:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.004086 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.004141 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.004151 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.004170 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.004182 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:19Z","lastTransitionTime":"2025-12-05T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.106868 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.106935 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.106952 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.106981 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.106997 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:19Z","lastTransitionTime":"2025-12-05T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.209760 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.209836 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.209848 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.209864 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.209876 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:19Z","lastTransitionTime":"2025-12-05T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.288146 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:19 crc kubenswrapper[4815]: E1205 09:08:19.288357 4815 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:08:19 crc kubenswrapper[4815]: E1205 09:08:19.288426 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs podName:010248a6-d07e-4e6b-97c8-bda1925d6adb nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.288405362 +0000 UTC m=+162.167012209 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs") pod "network-metrics-daemon-sslm2" (UID: "010248a6-d07e-4e6b-97c8-bda1925d6adb") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.312190 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.312227 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.312243 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.312272 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.312286 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:19Z","lastTransitionTime":"2025-12-05T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.414658 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.414699 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.414712 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.414728 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.414741 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:19Z","lastTransitionTime":"2025-12-05T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.417970 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.418005 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:19 crc kubenswrapper[4815]: E1205 09:08:19.418162 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.418179 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.418202 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:19 crc kubenswrapper[4815]: E1205 09:08:19.418311 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:19 crc kubenswrapper[4815]: E1205 09:08:19.418462 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:19 crc kubenswrapper[4815]: E1205 09:08:19.418675 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.517722 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.517764 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.517774 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.517788 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.517798 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:19Z","lastTransitionTime":"2025-12-05T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.620163 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.620225 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.620245 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.620269 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.620288 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:19Z","lastTransitionTime":"2025-12-05T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.722718 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.722767 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.722777 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.722792 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.722802 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:19Z","lastTransitionTime":"2025-12-05T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.825287 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.825321 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.825330 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.825379 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.825389 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:19Z","lastTransitionTime":"2025-12-05T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.928704 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.928779 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.928805 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.928838 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:19 crc kubenswrapper[4815]: I1205 09:08:19.928864 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:19Z","lastTransitionTime":"2025-12-05T09:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.032302 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.032352 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.032365 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.032383 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.032395 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:20Z","lastTransitionTime":"2025-12-05T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.134834 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.134915 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.134939 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.134971 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.134995 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:20Z","lastTransitionTime":"2025-12-05T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.238392 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.238455 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.238476 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.238541 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.238567 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:20Z","lastTransitionTime":"2025-12-05T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.340798 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.340889 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.340906 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.340927 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.340941 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:20Z","lastTransitionTime":"2025-12-05T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.443684 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.443746 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.443761 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.443783 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.443802 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:20Z","lastTransitionTime":"2025-12-05T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.546699 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.546773 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.546787 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.546807 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.546820 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:20Z","lastTransitionTime":"2025-12-05T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.649844 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.649895 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.649910 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.649935 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.649950 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:20Z","lastTransitionTime":"2025-12-05T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.752645 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.752708 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.752726 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.752748 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.752765 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:20Z","lastTransitionTime":"2025-12-05T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.855634 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.855706 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.855727 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.855759 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.855780 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:20Z","lastTransitionTime":"2025-12-05T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.959924 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.959975 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.959988 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.960002 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:20 crc kubenswrapper[4815]: I1205 09:08:20.960013 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:20Z","lastTransitionTime":"2025-12-05T09:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.062941 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.063005 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.063025 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.063052 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.063068 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:21Z","lastTransitionTime":"2025-12-05T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.166479 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.166613 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.166640 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.166676 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.166711 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:21Z","lastTransitionTime":"2025-12-05T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.269046 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.269076 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.269083 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.269097 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.269108 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:21Z","lastTransitionTime":"2025-12-05T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.372310 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.372377 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.372399 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.372424 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.372445 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:21Z","lastTransitionTime":"2025-12-05T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.418066 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.418184 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.418295 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.418645 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:21 crc kubenswrapper[4815]: E1205 09:08:21.418622 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:21 crc kubenswrapper[4815]: E1205 09:08:21.418780 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:21 crc kubenswrapper[4815]: E1205 09:08:21.418861 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:21 crc kubenswrapper[4815]: E1205 09:08:21.419274 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.437338 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d39b0996-638c-4ba5-93a2-3090496c33a8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f70f1427e22c5981c59a2e8031e275c9c9508c2f44189d66b858193497486f00\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a9d4a5a74073784e4555dbd172f06a17ad13eb63b6f3f8dee2e4b4f8f76a56d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f03537083edba3aaa6772ad4edea27b6b02e82939e9aff6618b50b0d4c617404\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d59cb3fb38a1161cc1f25cffd6038f2b1981f9fca1ffcee6ddd37e837a4fa946\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.454786 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.465311 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-ltlpb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"818e2dfa-a1d7-403d-a475-5437ece116cb\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://925c41a57d9dc421e12c462b16ee0adf4123ed66d3fc7d77766f8c09a176ecf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vvv88\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-ltlpb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.475329 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.475374 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.475386 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.475403 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.475414 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:21Z","lastTransitionTime":"2025-12-05T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.479205 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8c206810-e9ee-4480-8d79-8efcad288bb7\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3392a2d74d29f75e9794efe23bcb2808d70e32561692fbc759a06eb538bf7ad2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d5ce93b95790d1357e45b4eaafb3119331a642079a0a2a2730e725bf5bc351fc\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-05T09:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.491649 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fdf60992-6f16-4307-821e-f40e417fef65\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:00Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8095ac92a0340ac182451a96eb003f748176d6ca94de46d02ad705caee5d7948\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3a037ee90341ab082b8523235fd47661c15400baae0eaaa3f54340b6f5885962\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2d43960c9656d05594f1ecd47cd049e0ff80eea4abedc716f0c57aab93e031bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://815698fa132d64a49887e7988f69e7fbf24bce894d4565d8408f5a8dec09c786\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.505726 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d06b5523cf5f03a690f490494ed19b88c31a4562c66b763bf2552436e574fe11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.517116 4815 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0a874c44-831d-4f81-bf48-2025f5615574\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-05T09:07:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://60ec087cf467cd3283a677eca70a752481f1d16a1461691676f18803e7b3c512\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-05T09:07:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fxzq4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-05T09:07:02Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-4wq7d\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-05T09:08:21Z is after 2025-08-24T17:21:41Z" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.540315 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=80.540301538 podStartE2EDuration="1m20.540301538s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:21.540007489 +0000 UTC m=+100.418614336" watchObservedRunningTime="2025-12-05 09:08:21.540301538 +0000 UTC m=+100.418908365" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.576961 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.577010 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.577023 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.577045 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.577056 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:21Z","lastTransitionTime":"2025-12-05T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.627292 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-f9lkk" podStartSLOduration=80.627273059 podStartE2EDuration="1m20.627273059s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:21.598087902 +0000 UTC m=+100.476694749" watchObservedRunningTime="2025-12-05 09:08:21.627273059 +0000 UTC m=+100.505879896" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.641233 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-xd2x5" podStartSLOduration=80.641211433 podStartE2EDuration="1m20.641211433s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:21.640463991 +0000 UTC m=+100.519070828" watchObservedRunningTime="2025-12-05 09:08:21.641211433 +0000 UTC m=+100.519818270" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.679034 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.679082 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.679094 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.679111 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.679126 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:21Z","lastTransitionTime":"2025-12-05T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.715278 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=79.715258278 podStartE2EDuration="1m19.715258278s" podCreationTimestamp="2025-12-05 09:07:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:21.714822936 +0000 UTC m=+100.593429783" watchObservedRunningTime="2025-12-05 09:08:21.715258278 +0000 UTC m=+100.593865115" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.715876 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-625kt" podStartSLOduration=80.715869125 podStartE2EDuration="1m20.715869125s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:21.688662205 +0000 UTC m=+100.567269062" watchObservedRunningTime="2025-12-05 09:08:21.715869125 +0000 UTC m=+100.594475962" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.809569 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.809830 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.809905 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.809976 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.810035 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:21Z","lastTransitionTime":"2025-12-05T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.911893 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.912201 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.912282 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.912350 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:21 crc kubenswrapper[4815]: I1205 09:08:21.912412 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:21Z","lastTransitionTime":"2025-12-05T09:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.016056 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.016108 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.016123 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.016140 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.016153 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:22Z","lastTransitionTime":"2025-12-05T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.118630 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.118677 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.118689 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.118705 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.118717 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:22Z","lastTransitionTime":"2025-12-05T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.221548 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.221581 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.221591 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.221604 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.221613 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:22Z","lastTransitionTime":"2025-12-05T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.324803 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.324860 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.324879 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.324904 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.324921 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:22Z","lastTransitionTime":"2025-12-05T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.428015 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.428063 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.428081 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.428102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.428124 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:22Z","lastTransitionTime":"2025-12-05T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.532207 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.532277 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.532295 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.532323 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.532342 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:22Z","lastTransitionTime":"2025-12-05T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.635060 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.635185 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.635203 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.635227 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.635237 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:22Z","lastTransitionTime":"2025-12-05T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.739329 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.739406 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.739420 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.739447 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.739468 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:22Z","lastTransitionTime":"2025-12-05T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.842139 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.842190 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.842206 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.842229 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.842244 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:22Z","lastTransitionTime":"2025-12-05T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.945261 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.945335 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.945353 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.945856 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:22 crc kubenswrapper[4815]: I1205 09:08:22.945929 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:22Z","lastTransitionTime":"2025-12-05T09:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.048896 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.048941 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.048957 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.048979 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.048995 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:23Z","lastTransitionTime":"2025-12-05T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.152730 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.152789 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.152807 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.152833 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.152850 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:23Z","lastTransitionTime":"2025-12-05T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.255975 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.256034 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.256052 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.256075 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.256093 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:23Z","lastTransitionTime":"2025-12-05T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.362309 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.362385 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.362402 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.362424 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.362447 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:23Z","lastTransitionTime":"2025-12-05T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.417808 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.417855 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.417827 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:23 crc kubenswrapper[4815]: E1205 09:08:23.418024 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.418165 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:23 crc kubenswrapper[4815]: E1205 09:08:23.418414 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:23 crc kubenswrapper[4815]: E1205 09:08:23.418530 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:23 crc kubenswrapper[4815]: E1205 09:08:23.418626 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.465068 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.465123 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.465141 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.465164 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.465182 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:23Z","lastTransitionTime":"2025-12-05T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.568822 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.568908 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.568921 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.568943 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.568958 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:23Z","lastTransitionTime":"2025-12-05T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.671994 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.672031 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.672041 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.672055 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.672066 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:23Z","lastTransitionTime":"2025-12-05T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.774899 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.775017 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.775037 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.775061 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.775078 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:23Z","lastTransitionTime":"2025-12-05T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.877329 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.877370 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.877381 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.877397 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.877408 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:23Z","lastTransitionTime":"2025-12-05T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.980376 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.980420 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.980431 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.980452 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:23 crc kubenswrapper[4815]: I1205 09:08:23.980465 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:23Z","lastTransitionTime":"2025-12-05T09:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.082352 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.082395 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.082404 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.082418 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.082428 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:24Z","lastTransitionTime":"2025-12-05T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.186025 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.186144 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.186166 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.186193 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.186216 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:24Z","lastTransitionTime":"2025-12-05T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.290026 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.290097 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.290115 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.290140 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.290161 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:24Z","lastTransitionTime":"2025-12-05T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.393870 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.393941 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.393955 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.393981 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.393997 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:24Z","lastTransitionTime":"2025-12-05T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.419049 4815 scope.go:117] "RemoveContainer" containerID="c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca" Dec 05 09:08:24 crc kubenswrapper[4815]: E1205 09:08:24.419324 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.496549 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.496598 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.496610 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.496631 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.496643 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:24Z","lastTransitionTime":"2025-12-05T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.600018 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.600069 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.600102 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.600127 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.600141 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:24Z","lastTransitionTime":"2025-12-05T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.703040 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.703105 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.703130 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.703156 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.703174 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:24Z","lastTransitionTime":"2025-12-05T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.805746 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.805814 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.805831 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.805859 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.805876 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:24Z","lastTransitionTime":"2025-12-05T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.909164 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.909236 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.909264 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.909289 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:24 crc kubenswrapper[4815]: I1205 09:08:24.909306 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:24Z","lastTransitionTime":"2025-12-05T09:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.012746 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.012798 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.012811 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.012830 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.012842 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:25Z","lastTransitionTime":"2025-12-05T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.115856 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.115924 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.115950 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.115978 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.116002 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:25Z","lastTransitionTime":"2025-12-05T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.218898 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.218995 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.219014 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.219041 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.219062 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:25Z","lastTransitionTime":"2025-12-05T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.322079 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.322132 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.322145 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.322169 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.322183 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:25Z","lastTransitionTime":"2025-12-05T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.418684 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.418761 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.418723 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.418646 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:25 crc kubenswrapper[4815]: E1205 09:08:25.418893 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:25 crc kubenswrapper[4815]: E1205 09:08:25.418992 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:25 crc kubenswrapper[4815]: E1205 09:08:25.419065 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:25 crc kubenswrapper[4815]: E1205 09:08:25.419203 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.428711 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.428795 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.428820 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.428866 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.428896 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:25Z","lastTransitionTime":"2025-12-05T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.532363 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.532439 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.532459 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.532537 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.532561 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:25Z","lastTransitionTime":"2025-12-05T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.635646 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.635702 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.635715 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.635731 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.635744 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:25Z","lastTransitionTime":"2025-12-05T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.739932 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.739983 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.739994 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.740011 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.740024 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:25Z","lastTransitionTime":"2025-12-05T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.843282 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.843377 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.843394 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.843423 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.843438 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:25Z","lastTransitionTime":"2025-12-05T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.946830 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.946893 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.946914 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.946949 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:25 crc kubenswrapper[4815]: I1205 09:08:25.946971 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:25Z","lastTransitionTime":"2025-12-05T09:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.050389 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.050453 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.050463 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.050478 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.050503 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:26Z","lastTransitionTime":"2025-12-05T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.153611 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.153695 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.153714 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.153744 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.153770 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:26Z","lastTransitionTime":"2025-12-05T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.256929 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.256998 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.257016 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.257049 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.257071 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:26Z","lastTransitionTime":"2025-12-05T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.360120 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.360166 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.360177 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.360197 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.360210 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:26Z","lastTransitionTime":"2025-12-05T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.463418 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.463474 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.463504 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.463527 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.463548 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:26Z","lastTransitionTime":"2025-12-05T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.543805 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.543873 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.543885 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.543909 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.543924 4815 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-05T09:08:26Z","lastTransitionTime":"2025-12-05T09:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.597180 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-vrs47" podStartSLOduration=85.597154257 podStartE2EDuration="1m25.597154257s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:21.738544346 +0000 UTC m=+100.617151203" watchObservedRunningTime="2025-12-05 09:08:26.597154257 +0000 UTC m=+105.475761094" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.597786 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb"] Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.598562 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.601157 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.603753 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.603855 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.604062 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.612614 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podStartSLOduration=85.612596833 podStartE2EDuration="1m25.612596833s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:26.612392268 +0000 UTC m=+105.490999135" watchObservedRunningTime="2025-12-05 09:08:26.612596833 +0000 UTC m=+105.491203680" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.671971 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.672048 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.672079 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.672144 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.672190 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.692394 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-ltlpb" podStartSLOduration=86.692364491 podStartE2EDuration="1m26.692364491s" podCreationTimestamp="2025-12-05 09:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:26.691589028 +0000 UTC m=+105.570195865" watchObservedRunningTime="2025-12-05 09:08:26.692364491 +0000 UTC m=+105.570971328" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.702318 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=33.702294101 podStartE2EDuration="33.702294101s" podCreationTimestamp="2025-12-05 09:07:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:26.702128397 +0000 UTC m=+105.580735264" watchObservedRunningTime="2025-12-05 09:08:26.702294101 +0000 UTC m=+105.580900938" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.739355 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=56.739323629 podStartE2EDuration="56.739323629s" podCreationTimestamp="2025-12-05 09:07:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:26.738438844 +0000 UTC m=+105.617045701" watchObservedRunningTime="2025-12-05 09:08:26.739323629 +0000 UTC m=+105.617930476" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.740340 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=15.740333608 podStartE2EDuration="15.740333608s" podCreationTimestamp="2025-12-05 09:08:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:26.721203847 +0000 UTC m=+105.599810684" watchObservedRunningTime="2025-12-05 09:08:26.740333608 +0000 UTC m=+105.618940445" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.772964 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.773028 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.773083 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.773120 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.773164 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.773269 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.773316 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.774513 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-service-ca\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.780574 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.795785 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5bde293b-002a-40b6-9d0f-8bfbacadd7b5-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-lwxgb\" (UID: \"5bde293b-002a-40b6-9d0f-8bfbacadd7b5\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:26 crc kubenswrapper[4815]: I1205 09:08:26.923706 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" Dec 05 09:08:27 crc kubenswrapper[4815]: I1205 09:08:27.080730 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" event={"ID":"5bde293b-002a-40b6-9d0f-8bfbacadd7b5","Type":"ContainerStarted","Data":"bdbb8ba2988fbbf2ca6e01d00e15777e87d11a3e810454960d7941b291fa536f"} Dec 05 09:08:27 crc kubenswrapper[4815]: I1205 09:08:27.081445 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" event={"ID":"5bde293b-002a-40b6-9d0f-8bfbacadd7b5","Type":"ContainerStarted","Data":"0646a3b769937ab5b1e584260fdfa164995a8e43b42deda2744feadd9b0619c6"} Dec 05 09:08:27 crc kubenswrapper[4815]: I1205 09:08:27.094834 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-lwxgb" podStartSLOduration=87.094810537 podStartE2EDuration="1m27.094810537s" podCreationTimestamp="2025-12-05 09:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:08:27.094332434 +0000 UTC m=+105.972939281" watchObservedRunningTime="2025-12-05 09:08:27.094810537 +0000 UTC m=+105.973417384" Dec 05 09:08:27 crc kubenswrapper[4815]: I1205 09:08:27.418552 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:27 crc kubenswrapper[4815]: E1205 09:08:27.419183 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:27 crc kubenswrapper[4815]: I1205 09:08:27.418674 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:27 crc kubenswrapper[4815]: E1205 09:08:27.419394 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:27 crc kubenswrapper[4815]: I1205 09:08:27.418606 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:27 crc kubenswrapper[4815]: E1205 09:08:27.419594 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:27 crc kubenswrapper[4815]: I1205 09:08:27.418974 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:27 crc kubenswrapper[4815]: E1205 09:08:27.419819 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:29 crc kubenswrapper[4815]: I1205 09:08:29.418267 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:29 crc kubenswrapper[4815]: I1205 09:08:29.418358 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:29 crc kubenswrapper[4815]: I1205 09:08:29.418434 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:29 crc kubenswrapper[4815]: I1205 09:08:29.418260 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:29 crc kubenswrapper[4815]: E1205 09:08:29.418482 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:29 crc kubenswrapper[4815]: E1205 09:08:29.418567 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:29 crc kubenswrapper[4815]: E1205 09:08:29.418654 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:29 crc kubenswrapper[4815]: E1205 09:08:29.418761 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:31 crc kubenswrapper[4815]: I1205 09:08:31.418733 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:31 crc kubenswrapper[4815]: I1205 09:08:31.418867 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:31 crc kubenswrapper[4815]: I1205 09:08:31.420809 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:31 crc kubenswrapper[4815]: I1205 09:08:31.420872 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:31 crc kubenswrapper[4815]: E1205 09:08:31.421249 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:31 crc kubenswrapper[4815]: E1205 09:08:31.421096 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:31 crc kubenswrapper[4815]: E1205 09:08:31.421671 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:31 crc kubenswrapper[4815]: E1205 09:08:31.421869 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:33 crc kubenswrapper[4815]: I1205 09:08:33.418653 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:33 crc kubenswrapper[4815]: I1205 09:08:33.418722 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:33 crc kubenswrapper[4815]: I1205 09:08:33.418908 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:33 crc kubenswrapper[4815]: I1205 09:08:33.419119 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:33 crc kubenswrapper[4815]: E1205 09:08:33.419107 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:33 crc kubenswrapper[4815]: E1205 09:08:33.419277 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:33 crc kubenswrapper[4815]: E1205 09:08:33.419534 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:33 crc kubenswrapper[4815]: E1205 09:08:33.419677 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:35 crc kubenswrapper[4815]: I1205 09:08:35.109060 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-f9lkk_dff16910-bb36-4748-82bb-0f1d90da05b5/kube-multus/1.log" Dec 05 09:08:35 crc kubenswrapper[4815]: I1205 09:08:35.109640 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-f9lkk_dff16910-bb36-4748-82bb-0f1d90da05b5/kube-multus/0.log" Dec 05 09:08:35 crc kubenswrapper[4815]: I1205 09:08:35.109682 4815 generic.go:334] "Generic (PLEG): container finished" podID="dff16910-bb36-4748-82bb-0f1d90da05b5" containerID="93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b" exitCode=1 Dec 05 09:08:35 crc kubenswrapper[4815]: I1205 09:08:35.109712 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-f9lkk" event={"ID":"dff16910-bb36-4748-82bb-0f1d90da05b5","Type":"ContainerDied","Data":"93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b"} Dec 05 09:08:35 crc kubenswrapper[4815]: I1205 09:08:35.109749 4815 scope.go:117] "RemoveContainer" containerID="f543ccfc481f0b955b6fe010621efda6b8367e6cd25f83f6b1167ced5d0aaded" Dec 05 09:08:35 crc kubenswrapper[4815]: I1205 09:08:35.111211 4815 scope.go:117] "RemoveContainer" containerID="93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b" Dec 05 09:08:35 crc kubenswrapper[4815]: E1205 09:08:35.111494 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-f9lkk_openshift-multus(dff16910-bb36-4748-82bb-0f1d90da05b5)\"" pod="openshift-multus/multus-f9lkk" podUID="dff16910-bb36-4748-82bb-0f1d90da05b5" Dec 05 09:08:35 crc kubenswrapper[4815]: I1205 09:08:35.418314 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:35 crc kubenswrapper[4815]: E1205 09:08:35.418459 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:35 crc kubenswrapper[4815]: I1205 09:08:35.419251 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:35 crc kubenswrapper[4815]: I1205 09:08:35.419578 4815 scope.go:117] "RemoveContainer" containerID="c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca" Dec 05 09:08:35 crc kubenswrapper[4815]: E1205 09:08:35.419778 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" Dec 05 09:08:35 crc kubenswrapper[4815]: E1205 09:08:35.420001 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:35 crc kubenswrapper[4815]: I1205 09:08:35.420144 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:35 crc kubenswrapper[4815]: E1205 09:08:35.420546 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:35 crc kubenswrapper[4815]: I1205 09:08:35.420167 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:35 crc kubenswrapper[4815]: E1205 09:08:35.421115 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:36 crc kubenswrapper[4815]: I1205 09:08:36.114365 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-f9lkk_dff16910-bb36-4748-82bb-0f1d90da05b5/kube-multus/1.log" Dec 05 09:08:37 crc kubenswrapper[4815]: I1205 09:08:37.418009 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:37 crc kubenswrapper[4815]: I1205 09:08:37.418023 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:37 crc kubenswrapper[4815]: I1205 09:08:37.418063 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:37 crc kubenswrapper[4815]: I1205 09:08:37.418123 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:37 crc kubenswrapper[4815]: E1205 09:08:37.418294 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:37 crc kubenswrapper[4815]: E1205 09:08:37.418789 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:37 crc kubenswrapper[4815]: E1205 09:08:37.418949 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:37 crc kubenswrapper[4815]: E1205 09:08:37.419037 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:39 crc kubenswrapper[4815]: I1205 09:08:39.418283 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:39 crc kubenswrapper[4815]: E1205 09:08:39.418412 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:39 crc kubenswrapper[4815]: I1205 09:08:39.418284 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:39 crc kubenswrapper[4815]: I1205 09:08:39.418572 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:39 crc kubenswrapper[4815]: E1205 09:08:39.418688 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:39 crc kubenswrapper[4815]: E1205 09:08:39.418759 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:39 crc kubenswrapper[4815]: I1205 09:08:39.420013 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:39 crc kubenswrapper[4815]: E1205 09:08:39.420569 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:41 crc kubenswrapper[4815]: E1205 09:08:41.400937 4815 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 05 09:08:41 crc kubenswrapper[4815]: I1205 09:08:41.418212 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:41 crc kubenswrapper[4815]: I1205 09:08:41.418241 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:41 crc kubenswrapper[4815]: I1205 09:08:41.418338 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:41 crc kubenswrapper[4815]: E1205 09:08:41.420277 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:41 crc kubenswrapper[4815]: I1205 09:08:41.420313 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:41 crc kubenswrapper[4815]: E1205 09:08:41.420432 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:41 crc kubenswrapper[4815]: E1205 09:08:41.420602 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:41 crc kubenswrapper[4815]: E1205 09:08:41.420832 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:41 crc kubenswrapper[4815]: E1205 09:08:41.577838 4815 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 09:08:43 crc kubenswrapper[4815]: I1205 09:08:43.418201 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:43 crc kubenswrapper[4815]: I1205 09:08:43.418201 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:43 crc kubenswrapper[4815]: I1205 09:08:43.418266 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:43 crc kubenswrapper[4815]: I1205 09:08:43.418381 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:43 crc kubenswrapper[4815]: E1205 09:08:43.418565 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:43 crc kubenswrapper[4815]: E1205 09:08:43.418956 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:43 crc kubenswrapper[4815]: E1205 09:08:43.419131 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:43 crc kubenswrapper[4815]: E1205 09:08:43.419270 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:45 crc kubenswrapper[4815]: I1205 09:08:45.418259 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:45 crc kubenswrapper[4815]: I1205 09:08:45.418318 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:45 crc kubenswrapper[4815]: E1205 09:08:45.418376 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:45 crc kubenswrapper[4815]: I1205 09:08:45.418397 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:45 crc kubenswrapper[4815]: I1205 09:08:45.418262 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:45 crc kubenswrapper[4815]: E1205 09:08:45.418572 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:45 crc kubenswrapper[4815]: E1205 09:08:45.418697 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:45 crc kubenswrapper[4815]: E1205 09:08:45.418792 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:46 crc kubenswrapper[4815]: I1205 09:08:46.419151 4815 scope.go:117] "RemoveContainer" containerID="93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b" Dec 05 09:08:46 crc kubenswrapper[4815]: E1205 09:08:46.579106 4815 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 09:08:47 crc kubenswrapper[4815]: I1205 09:08:47.153572 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-f9lkk_dff16910-bb36-4748-82bb-0f1d90da05b5/kube-multus/1.log" Dec 05 09:08:47 crc kubenswrapper[4815]: I1205 09:08:47.153631 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-f9lkk" event={"ID":"dff16910-bb36-4748-82bb-0f1d90da05b5","Type":"ContainerStarted","Data":"956f4463e9520b777cb14ec518d485199408c873c9afec4b38e56f0e0ec02fbf"} Dec 05 09:08:47 crc kubenswrapper[4815]: I1205 09:08:47.418122 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:47 crc kubenswrapper[4815]: I1205 09:08:47.418179 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:47 crc kubenswrapper[4815]: I1205 09:08:47.418143 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:47 crc kubenswrapper[4815]: E1205 09:08:47.418266 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:47 crc kubenswrapper[4815]: I1205 09:08:47.418122 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:47 crc kubenswrapper[4815]: E1205 09:08:47.418407 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:47 crc kubenswrapper[4815]: E1205 09:08:47.418815 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:47 crc kubenswrapper[4815]: E1205 09:08:47.418945 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:47 crc kubenswrapper[4815]: I1205 09:08:47.419116 4815 scope.go:117] "RemoveContainer" containerID="c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca" Dec 05 09:08:47 crc kubenswrapper[4815]: E1205 09:08:47.419295 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-pzlxw_openshift-ovn-kubernetes(e930cfad-5770-4db6-826b-353554e2f23c)\"" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" Dec 05 09:08:49 crc kubenswrapper[4815]: I1205 09:08:49.418525 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:49 crc kubenswrapper[4815]: I1205 09:08:49.418587 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:49 crc kubenswrapper[4815]: I1205 09:08:49.418587 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:49 crc kubenswrapper[4815]: E1205 09:08:49.418648 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:49 crc kubenswrapper[4815]: E1205 09:08:49.418757 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:49 crc kubenswrapper[4815]: I1205 09:08:49.418806 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:49 crc kubenswrapper[4815]: E1205 09:08:49.418828 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:49 crc kubenswrapper[4815]: E1205 09:08:49.418972 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:51 crc kubenswrapper[4815]: I1205 09:08:51.417847 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:51 crc kubenswrapper[4815]: I1205 09:08:51.417862 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:51 crc kubenswrapper[4815]: I1205 09:08:51.417907 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:51 crc kubenswrapper[4815]: I1205 09:08:51.417929 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:51 crc kubenswrapper[4815]: E1205 09:08:51.419818 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:51 crc kubenswrapper[4815]: E1205 09:08:51.419976 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:51 crc kubenswrapper[4815]: E1205 09:08:51.420081 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:51 crc kubenswrapper[4815]: E1205 09:08:51.420366 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:51 crc kubenswrapper[4815]: E1205 09:08:51.580605 4815 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 09:08:53 crc kubenswrapper[4815]: I1205 09:08:53.418377 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:53 crc kubenswrapper[4815]: I1205 09:08:53.418529 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:53 crc kubenswrapper[4815]: E1205 09:08:53.418716 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:53 crc kubenswrapper[4815]: I1205 09:08:53.418769 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:53 crc kubenswrapper[4815]: I1205 09:08:53.418820 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:53 crc kubenswrapper[4815]: E1205 09:08:53.418876 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:53 crc kubenswrapper[4815]: E1205 09:08:53.418976 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:53 crc kubenswrapper[4815]: E1205 09:08:53.419066 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:55 crc kubenswrapper[4815]: I1205 09:08:55.418667 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:55 crc kubenswrapper[4815]: I1205 09:08:55.418767 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:55 crc kubenswrapper[4815]: I1205 09:08:55.418683 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:55 crc kubenswrapper[4815]: E1205 09:08:55.418871 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:55 crc kubenswrapper[4815]: I1205 09:08:55.418956 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:55 crc kubenswrapper[4815]: E1205 09:08:55.419057 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:55 crc kubenswrapper[4815]: E1205 09:08:55.419198 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:55 crc kubenswrapper[4815]: E1205 09:08:55.419375 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:56 crc kubenswrapper[4815]: E1205 09:08:56.582209 4815 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 09:08:57 crc kubenswrapper[4815]: I1205 09:08:57.418126 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:57 crc kubenswrapper[4815]: I1205 09:08:57.418188 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:57 crc kubenswrapper[4815]: I1205 09:08:57.418189 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:57 crc kubenswrapper[4815]: E1205 09:08:57.418325 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:57 crc kubenswrapper[4815]: E1205 09:08:57.418465 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:57 crc kubenswrapper[4815]: E1205 09:08:57.418743 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:08:57 crc kubenswrapper[4815]: I1205 09:08:57.419083 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:57 crc kubenswrapper[4815]: E1205 09:08:57.419228 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:59 crc kubenswrapper[4815]: I1205 09:08:59.418744 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:08:59 crc kubenswrapper[4815]: I1205 09:08:59.418833 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:08:59 crc kubenswrapper[4815]: I1205 09:08:59.418744 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:08:59 crc kubenswrapper[4815]: E1205 09:08:59.418964 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:08:59 crc kubenswrapper[4815]: I1205 09:08:59.419136 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:08:59 crc kubenswrapper[4815]: E1205 09:08:59.419240 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:08:59 crc kubenswrapper[4815]: E1205 09:08:59.419930 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:08:59 crc kubenswrapper[4815]: E1205 09:08:59.419467 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:09:01 crc kubenswrapper[4815]: I1205 09:09:01.418560 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:09:01 crc kubenswrapper[4815]: I1205 09:09:01.418582 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:09:01 crc kubenswrapper[4815]: I1205 09:09:01.418679 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:09:01 crc kubenswrapper[4815]: I1205 09:09:01.420038 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:09:01 crc kubenswrapper[4815]: E1205 09:09:01.420044 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:09:01 crc kubenswrapper[4815]: E1205 09:09:01.420157 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:09:01 crc kubenswrapper[4815]: E1205 09:09:01.420251 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:09:01 crc kubenswrapper[4815]: E1205 09:09:01.420307 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:09:01 crc kubenswrapper[4815]: E1205 09:09:01.583389 4815 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 05 09:09:02 crc kubenswrapper[4815]: I1205 09:09:02.418797 4815 scope.go:117] "RemoveContainer" containerID="c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca" Dec 05 09:09:03 crc kubenswrapper[4815]: I1205 09:09:03.211545 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/3.log" Dec 05 09:09:03 crc kubenswrapper[4815]: I1205 09:09:03.215903 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerStarted","Data":"5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76"} Dec 05 09:09:03 crc kubenswrapper[4815]: I1205 09:09:03.216286 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:09:03 crc kubenswrapper[4815]: I1205 09:09:03.239760 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podStartSLOduration=122.239743477 podStartE2EDuration="2m2.239743477s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:03.238952244 +0000 UTC m=+142.117559101" watchObservedRunningTime="2025-12-05 09:09:03.239743477 +0000 UTC m=+142.118350304" Dec 05 09:09:03 crc kubenswrapper[4815]: I1205 09:09:03.418291 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:09:03 crc kubenswrapper[4815]: I1205 09:09:03.418335 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:09:03 crc kubenswrapper[4815]: E1205 09:09:03.418407 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:09:03 crc kubenswrapper[4815]: I1205 09:09:03.418447 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:09:03 crc kubenswrapper[4815]: E1205 09:09:03.418540 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:09:03 crc kubenswrapper[4815]: E1205 09:09:03.418591 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:09:03 crc kubenswrapper[4815]: I1205 09:09:03.418279 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:09:03 crc kubenswrapper[4815]: E1205 09:09:03.418695 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:09:03 crc kubenswrapper[4815]: I1205 09:09:03.548813 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-sslm2"] Dec 05 09:09:04 crc kubenswrapper[4815]: I1205 09:09:04.219850 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:09:04 crc kubenswrapper[4815]: E1205 09:09:04.219980 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:09:05 crc kubenswrapper[4815]: I1205 09:09:05.419773 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:09:05 crc kubenswrapper[4815]: I1205 09:09:05.419808 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:09:05 crc kubenswrapper[4815]: I1205 09:09:05.419890 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:09:05 crc kubenswrapper[4815]: E1205 09:09:05.419995 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 05 09:09:05 crc kubenswrapper[4815]: I1205 09:09:05.420163 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:09:05 crc kubenswrapper[4815]: E1205 09:09:05.420205 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 05 09:09:05 crc kubenswrapper[4815]: E1205 09:09:05.420316 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-sslm2" podUID="010248a6-d07e-4e6b-97c8-bda1925d6adb" Dec 05 09:09:05 crc kubenswrapper[4815]: E1205 09:09:05.420416 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.418366 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.418375 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.418417 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.418545 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.422921 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.423197 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.423882 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.424479 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.425011 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.427680 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.711242 4815 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.779708 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tw877"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.780626 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.783640 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.786276 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.787090 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.787856 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.789110 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.789919 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.800027 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.800835 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.802464 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2nvwc"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.803324 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.803859 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-ssjtf"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.804315 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ssjtf" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.806999 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.807899 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.808007 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-d2j7q"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.808668 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.809565 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h66pq"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.810310 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.811093 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.811826 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.812394 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wjxhv"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.812899 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.826952 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.827029 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.827288 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.827146 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.829700 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.829905 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.829950 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.830111 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.830186 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.830288 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.830361 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.832086 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.832315 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.832470 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.832642 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.832822 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.833003 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.834200 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.834558 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.834627 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ltjzv"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.834731 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.835115 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.835158 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.852298 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.853006 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.853127 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.853170 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.853190 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.853136 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.853024 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.853100 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.853416 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.853017 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.853606 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.853694 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.855184 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.855659 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.855776 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.857728 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.858002 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.858187 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.858376 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fm9ss"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.858826 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-skb9s"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.859253 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.859302 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4mvlm"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.859531 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.859655 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860040 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860144 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860231 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860342 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860393 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860446 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860553 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860573 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860678 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860697 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860795 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860830 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860958 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.861055 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.860958 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.861279 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.864377 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.865530 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.867482 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fvz7z"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.868693 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.869756 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.870697 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.870877 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.872823 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.874309 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.876186 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.876616 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.878449 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.879658 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.879873 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.882195 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.882560 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.887638 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.894813 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.895311 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922017 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/99bb7807-d522-499e-863e-7be3e10ac6ed-encryption-config\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922064 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-config\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922085 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922112 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8djn\" (UniqueName: \"kubernetes.io/projected/99bb7807-d522-499e-863e-7be3e10ac6ed-kube-api-access-l8djn\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922133 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a34781cf-0b80-4615-913a-8908ca380ef2-encryption-config\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922157 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-image-import-ca\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922193 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a34781cf-0b80-4615-913a-8908ca380ef2-serving-cert\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922214 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99bb7807-d522-499e-863e-7be3e10ac6ed-etcd-client\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922237 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a34781cf-0b80-4615-913a-8908ca380ef2-etcd-client\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922258 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a34781cf-0b80-4615-913a-8908ca380ef2-audit-dir\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922280 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/afd36151-d969-4642-b6c4-c6b3b5f65476-images\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922332 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6rzb\" (UniqueName: \"kubernetes.io/projected/a34781cf-0b80-4615-913a-8908ca380ef2-kube-api-access-h6rzb\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922360 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-etcd-serving-ca\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922385 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afd36151-d969-4642-b6c4-c6b3b5f65476-config\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922420 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkmkm\" (UniqueName: \"kubernetes.io/projected/5655913f-6305-4b7c-bfef-2795e06e0a29-kube-api-access-kkmkm\") pod \"downloads-7954f5f757-ssjtf\" (UID: \"5655913f-6305-4b7c-bfef-2795e06e0a29\") " pod="openshift-console/downloads-7954f5f757-ssjtf" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922442 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a34781cf-0b80-4615-913a-8908ca380ef2-audit-policies\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.922464 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/99bb7807-d522-499e-863e-7be3e10ac6ed-node-pullsecrets\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.924938 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/afd36151-d969-4642-b6c4-c6b3b5f65476-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.925029 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vc2th\" (UniqueName: \"kubernetes.io/projected/afd36151-d969-4642-b6c4-c6b3b5f65476-kube-api-access-vc2th\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.925088 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a34781cf-0b80-4615-913a-8908ca380ef2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.925164 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99bb7807-d522-499e-863e-7be3e10ac6ed-serving-cert\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.925197 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99bb7807-d522-499e-863e-7be3e10ac6ed-audit-dir\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.925232 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-audit\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.925265 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a34781cf-0b80-4615-913a-8908ca380ef2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.929689 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.978212 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.978530 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.979229 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.981391 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjd8"] Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.982305 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.988773 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.989463 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.989939 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.990237 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.990336 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 09:09:07 crc kubenswrapper[4815]: I1205 09:09:07.990424 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.025622 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.026094 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.026628 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.026911 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.027037 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.027702 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028542 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1615b53d-0e1b-4d6a-a369-1a614711721c-config\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028609 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6rzb\" (UniqueName: \"kubernetes.io/projected/a34781cf-0b80-4615-913a-8908ca380ef2-kube-api-access-h6rzb\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028630 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2q5t\" (UniqueName: \"kubernetes.io/projected/97c971b0-67a2-456b-ae6b-eb4f2a69630d-kube-api-access-n2q5t\") pod \"cluster-samples-operator-665b6dd947-2w9rk\" (UID: \"97c971b0-67a2-456b-ae6b-eb4f2a69630d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028651 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-etcd-serving-ca\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028661 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028666 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028779 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028828 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028852 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028667 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1615b53d-0e1b-4d6a-a369-1a614711721c-machine-approver-tls\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028943 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.028941 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afd36151-d969-4642-b6c4-c6b3b5f65476-config\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029102 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-config\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029134 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029156 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkmkm\" (UniqueName: \"kubernetes.io/projected/5655913f-6305-4b7c-bfef-2795e06e0a29-kube-api-access-kkmkm\") pod \"downloads-7954f5f757-ssjtf\" (UID: \"5655913f-6305-4b7c-bfef-2795e06e0a29\") " pod="openshift-console/downloads-7954f5f757-ssjtf" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029175 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a34781cf-0b80-4615-913a-8908ca380ef2-audit-policies\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029191 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/99bb7807-d522-499e-863e-7be3e10ac6ed-node-pullsecrets\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029205 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vd7n\" (UniqueName: \"kubernetes.io/projected/225d33d1-7da4-4937-a843-7594d7549454-kube-api-access-4vd7n\") pod \"dns-operator-744455d44c-ltjzv\" (UID: \"225d33d1-7da4-4937-a843-7594d7549454\") " pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029226 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/afd36151-d969-4642-b6c4-c6b3b5f65476-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029246 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vc2th\" (UniqueName: \"kubernetes.io/projected/afd36151-d969-4642-b6c4-c6b3b5f65476-kube-api-access-vc2th\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029269 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a34781cf-0b80-4615-913a-8908ca380ef2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029288 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99bb7807-d522-499e-863e-7be3e10ac6ed-serving-cert\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029302 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1615b53d-0e1b-4d6a-a369-1a614711721c-auth-proxy-config\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029317 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-service-ca-bundle\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029335 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99bb7807-d522-499e-863e-7be3e10ac6ed-audit-dir\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029349 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-audit\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029366 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a34781cf-0b80-4615-913a-8908ca380ef2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029381 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-serving-cert\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029412 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/99bb7807-d522-499e-863e-7be3e10ac6ed-encryption-config\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029428 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrxtr\" (UniqueName: \"kubernetes.io/projected/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-kube-api-access-wrxtr\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029447 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-config\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029461 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029479 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8djn\" (UniqueName: \"kubernetes.io/projected/99bb7807-d522-499e-863e-7be3e10ac6ed-kube-api-access-l8djn\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029520 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/97c971b0-67a2-456b-ae6b-eb4f2a69630d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2w9rk\" (UID: \"97c971b0-67a2-456b-ae6b-eb4f2a69630d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029538 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a34781cf-0b80-4615-913a-8908ca380ef2-encryption-config\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029557 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-image-import-ca\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029585 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a34781cf-0b80-4615-913a-8908ca380ef2-serving-cert\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029604 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99bb7807-d522-499e-863e-7be3e10ac6ed-etcd-client\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029622 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a34781cf-0b80-4615-913a-8908ca380ef2-etcd-client\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029655 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/225d33d1-7da4-4937-a843-7594d7549454-metrics-tls\") pod \"dns-operator-744455d44c-ltjzv\" (UID: \"225d33d1-7da4-4937-a843-7594d7549454\") " pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029674 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a34781cf-0b80-4615-913a-8908ca380ef2-audit-dir\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029690 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/afd36151-d969-4642-b6c4-c6b3b5f65476-images\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029713 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vsrn\" (UniqueName: \"kubernetes.io/projected/1615b53d-0e1b-4d6a-a369-1a614711721c-kube-api-access-2vsrn\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029713 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afd36151-d969-4642-b6c4-c6b3b5f65476-config\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.029993 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.030451 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a34781cf-0b80-4615-913a-8908ca380ef2-audit-policies\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.030478 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/99bb7807-d522-499e-863e-7be3e10ac6ed-audit-dir\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.030920 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-audit\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.031057 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/99bb7807-d522-499e-863e-7be3e10ac6ed-node-pullsecrets\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.031456 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a34781cf-0b80-4615-913a-8908ca380ef2-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.034393 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.034994 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a34781cf-0b80-4615-913a-8908ca380ef2-audit-dir\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.036130 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.037739 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.038093 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.038294 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.038784 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.038843 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.040098 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/afd36151-d969-4642-b6c4-c6b3b5f65476-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.040981 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/afd36151-d969-4642-b6c4-c6b3b5f65476-images\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.042109 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/a34781cf-0b80-4615-913a-8908ca380ef2-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.042463 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.042524 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-etcd-serving-ca\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.042687 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-v2lvl"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.042971 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-config\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.043755 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.044369 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.045853 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.046235 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.051910 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.053174 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.053627 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.054102 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.054441 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.054741 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.056857 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-trusted-ca-bundle\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.056955 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.057459 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.061469 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.062112 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.063892 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/a34781cf-0b80-4615-913a-8908ca380ef2-encryption-config\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.064742 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-cgjsg"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.065336 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.065422 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.065814 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.066911 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a34781cf-0b80-4615-913a-8908ca380ef2-etcd-client\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.067975 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.068443 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.068853 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-47vsq"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.068911 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a34781cf-0b80-4615-913a-8908ca380ef2-serving-cert\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.069030 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.069535 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.069795 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.070543 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgk4c"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.070960 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.071124 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4jpx4"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.071378 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.071520 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.071559 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.071727 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.071928 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.072039 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/99bb7807-d522-499e-863e-7be3e10ac6ed-serving-cert\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.072531 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.072674 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/99bb7807-d522-499e-863e-7be3e10ac6ed-image-import-ca\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.072777 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.073732 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/99bb7807-d522-499e-863e-7be3e10ac6ed-etcd-client\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.073834 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.074030 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.074211 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.075788 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/99bb7807-d522-499e-863e-7be3e10ac6ed-encryption-config\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.075829 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.076286 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6rzb\" (UniqueName: \"kubernetes.io/projected/a34781cf-0b80-4615-913a-8908ca380ef2-kube-api-access-h6rzb\") pod \"apiserver-7bbb656c7d-fkstw\" (UID: \"a34781cf-0b80-4615-913a-8908ca380ef2\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.076349 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.077693 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.092317 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkmkm\" (UniqueName: \"kubernetes.io/projected/5655913f-6305-4b7c-bfef-2795e06e0a29-kube-api-access-kkmkm\") pod \"downloads-7954f5f757-ssjtf\" (UID: \"5655913f-6305-4b7c-bfef-2795e06e0a29\") " pod="openshift-console/downloads-7954f5f757-ssjtf" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.093247 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.093676 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2nvwc"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.093781 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.093838 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.099585 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.116415 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vc2th\" (UniqueName: \"kubernetes.io/projected/afd36151-d969-4642-b6c4-c6b3b5f65476-kube-api-access-vc2th\") pod \"machine-api-operator-5694c8668f-tw877\" (UID: \"afd36151-d969-4642-b6c4-c6b3b5f65476\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.116581 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.116654 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ssjtf"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.116694 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tw877"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.117878 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.128390 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-ssjtf" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.129642 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130484 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/225d33d1-7da4-4937-a843-7594d7549454-metrics-tls\") pod \"dns-operator-744455d44c-ltjzv\" (UID: \"225d33d1-7da4-4937-a843-7594d7549454\") " pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130534 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vsrn\" (UniqueName: \"kubernetes.io/projected/1615b53d-0e1b-4d6a-a369-1a614711721c-kube-api-access-2vsrn\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130562 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1615b53d-0e1b-4d6a-a369-1a614711721c-config\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130590 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2q5t\" (UniqueName: \"kubernetes.io/projected/97c971b0-67a2-456b-ae6b-eb4f2a69630d-kube-api-access-n2q5t\") pod \"cluster-samples-operator-665b6dd947-2w9rk\" (UID: \"97c971b0-67a2-456b-ae6b-eb4f2a69630d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130609 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1615b53d-0e1b-4d6a-a369-1a614711721c-machine-approver-tls\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130652 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-config\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130674 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130691 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vd7n\" (UniqueName: \"kubernetes.io/projected/225d33d1-7da4-4937-a843-7594d7549454-kube-api-access-4vd7n\") pod \"dns-operator-744455d44c-ltjzv\" (UID: \"225d33d1-7da4-4937-a843-7594d7549454\") " pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130726 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1615b53d-0e1b-4d6a-a369-1a614711721c-auth-proxy-config\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130744 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-service-ca-bundle\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130796 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-serving-cert\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130814 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrxtr\" (UniqueName: \"kubernetes.io/projected/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-kube-api-access-wrxtr\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.130848 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/97c971b0-67a2-456b-ae6b-eb4f2a69630d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2w9rk\" (UID: \"97c971b0-67a2-456b-ae6b-eb4f2a69630d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.133324 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-config\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.133660 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1615b53d-0e1b-4d6a-a369-1a614711721c-config\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.134091 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.134316 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.134678 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-service-ca-bundle\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.138582 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/97c971b0-67a2-456b-ae6b-eb4f2a69630d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-2w9rk\" (UID: \"97c971b0-67a2-456b-ae6b-eb4f2a69630d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.138690 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wjxhv"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.139166 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/1615b53d-0e1b-4d6a-a369-1a614711721c-machine-approver-tls\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.142010 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-kj46k"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.142385 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-serving-cert\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.143596 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/225d33d1-7da4-4937-a843-7594d7549454-metrics-tls\") pod \"dns-operator-744455d44c-ltjzv\" (UID: \"225d33d1-7da4-4937-a843-7594d7549454\") " pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.144089 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8djn\" (UniqueName: \"kubernetes.io/projected/99bb7807-d522-499e-863e-7be3e10ac6ed-kube-api-access-l8djn\") pod \"apiserver-76f77b778f-2nvwc\" (UID: \"99bb7807-d522-499e-863e-7be3e10ac6ed\") " pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.144302 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.144521 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1615b53d-0e1b-4d6a-a369-1a614711721c-auth-proxy-config\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.146650 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kj46k" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.149966 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-d2j7q"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.150005 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ltjzv"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.150038 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fvz7z"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.152117 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.152551 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.153582 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fm9ss"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.156986 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.157122 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-skb9s"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.158549 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h66pq"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.159709 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.159834 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.161768 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4mvlm"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.163044 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.163265 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-d5gt5"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.164271 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.164641 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.166140 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.167431 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.167792 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.169449 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-rhtl7"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.170169 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjd8"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.170359 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rhtl7" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.172435 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.174046 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-54d4p"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.175115 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.176723 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.179199 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.180612 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.181542 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-47vsq"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.182548 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgk4c"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.183068 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.183897 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-cgjsg"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.184899 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.185898 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.187475 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.188564 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.189927 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.191072 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-54d4p"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.192167 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4jpx4"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.193235 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-d5gt5"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.194306 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.195530 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.196713 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rhtl7"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.203594 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.222596 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.242686 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.266536 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.346924 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.347097 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.354145 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.362767 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.363144 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.383652 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.403526 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.423780 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.442653 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.484303 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.504259 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.512259 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.526063 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.585534 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.585631 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.585781 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.606515 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.623714 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.642384 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.662824 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.687895 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.703744 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.723403 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.745072 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.764673 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.782111 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.802806 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.822681 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.843203 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.863648 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.896859 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.904714 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.924922 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.945072 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.965125 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.985008 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.998517 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw"] Dec 05 09:09:08 crc kubenswrapper[4815]: I1205 09:09:08.998554 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-tw877"] Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.003183 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.025096 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.030614 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-ssjtf"] Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.042483 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.055613 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-2nvwc"] Dec 05 09:09:09 crc kubenswrapper[4815]: W1205 09:09:09.061340 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99bb7807_d522_499e_863e_7be3e10ac6ed.slice/crio-c20bdb6c753ec5cb8cab52b7652fbbcbd32e9d9ff91079f39c06804f01a56500 WatchSource:0}: Error finding container c20bdb6c753ec5cb8cab52b7652fbbcbd32e9d9ff91079f39c06804f01a56500: Status 404 returned error can't find the container with id c20bdb6c753ec5cb8cab52b7652fbbcbd32e9d9ff91079f39c06804f01a56500 Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.062668 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.081200 4815 request.go:700] Waited for 1.015181792s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/secrets?fieldSelector=metadata.name%3Dcontrol-plane-machine-set-operator-dockercfg-k9rxt&limit=500&resourceVersion=0 Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.082953 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.103546 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.122949 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.142657 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.162668 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.183114 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.203909 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.223089 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.242519 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.252480 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ssjtf" event={"ID":"5655913f-6305-4b7c-bfef-2795e06e0a29","Type":"ContainerStarted","Data":"65fc987e195d1995d818656d0c138c3c22672210a737ab622cdee7143f09f698"} Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.252635 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ssjtf" event={"ID":"5655913f-6305-4b7c-bfef-2795e06e0a29","Type":"ContainerStarted","Data":"11ae327c1feeec3060902bf03dc9e641fd0b43fd3a8e0f11c6413b8b7e256bcf"} Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.252896 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-ssjtf" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.254374 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" event={"ID":"afd36151-d969-4642-b6c4-c6b3b5f65476","Type":"ContainerStarted","Data":"455af9b291ce89f8fd3b979af2f40ad61d3fde54c2db474ed9d0a08addf88d91"} Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.254481 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" event={"ID":"afd36151-d969-4642-b6c4-c6b3b5f65476","Type":"ContainerStarted","Data":"17e53550c9f84e6cde0ef1c835dddb3f5f3fc4e3fc8e3efa7e21c3535e31999f"} Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.255974 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" event={"ID":"99bb7807-d522-499e-863e-7be3e10ac6ed","Type":"ContainerStarted","Data":"c20bdb6c753ec5cb8cab52b7652fbbcbd32e9d9ff91079f39c06804f01a56500"} Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.257303 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" event={"ID":"a34781cf-0b80-4615-913a-8908ca380ef2","Type":"ContainerStarted","Data":"2a4168c1eb7da717abe2eb85302699fba851d6e0c132109fa8e55f81afe73d20"} Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.257799 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.257907 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.262658 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.283097 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.303162 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.322700 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.342707 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.367371 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.383383 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.396015 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:09 crc kubenswrapper[4815]: E1205 09:09:09.396356 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:11:11.396335415 +0000 UTC m=+270.274942262 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.403113 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.422872 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.442406 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.470014 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.482625 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.497532 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.497580 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.497604 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.497636 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.499224 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.504897 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.507377 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.511574 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.511964 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.524156 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.549093 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.562006 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.564070 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.578988 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599019 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6bs4\" (UniqueName: \"kubernetes.io/projected/036169eb-913d-470b-b749-6583deb8f396-kube-api-access-r6bs4\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599064 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-service-ca\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599083 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-config\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599113 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-etcd-service-ca\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599132 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6v6w\" (UniqueName: \"kubernetes.io/projected/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-kube-api-access-d6v6w\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599149 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-trusted-ca\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599165 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85tbf\" (UniqueName: \"kubernetes.io/projected/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-kube-api-access-85tbf\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599185 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a20f4d0e-9b06-4bc0-a532-a104392a18dd-config\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599198 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-bound-sa-token\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599214 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2211eb63-8b44-48c6-9cca-c032127ddb95-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kvstg\" (UID: \"2211eb63-8b44-48c6-9cca-c032127ddb95\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599229 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-serving-cert\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599247 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599260 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-oauth-config\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599273 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-etcd-client\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599289 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-trusted-ca\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599305 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2211eb63-8b44-48c6-9cca-c032127ddb95-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kvstg\" (UID: \"2211eb63-8b44-48c6-9cca-c032127ddb95\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599327 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a6ab936d-3cc8-4c09-869a-50783298666d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599341 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-config\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599357 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrd58\" (UniqueName: \"kubernetes.io/projected/2211eb63-8b44-48c6-9cca-c032127ddb95-kube-api-access-hrd58\") pod \"openshift-controller-manager-operator-756b6f6bc6-kvstg\" (UID: \"2211eb63-8b44-48c6-9cca-c032127ddb95\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599372 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/be45782f-d05a-4f70-8d05-3556a37a9698-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599389 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/be45782f-d05a-4f70-8d05-3556a37a9698-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599405 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2znbq\" (UniqueName: \"kubernetes.io/projected/be45782f-d05a-4f70-8d05-3556a37a9698-kube-api-access-2znbq\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599423 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be45782f-d05a-4f70-8d05-3556a37a9698-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599438 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9190adfa-bf4f-421b-a2d9-21bb94b94334-available-featuregates\") pod \"openshift-config-operator-7777fb866f-skb9s\" (UID: \"9190adfa-bf4f-421b-a2d9-21bb94b94334\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599452 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-config\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599473 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cf7wz\" (UniqueName: \"kubernetes.io/projected/a20f4d0e-9b06-4bc0-a532-a104392a18dd-kube-api-access-cf7wz\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599504 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-client-ca\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599523 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-config\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599544 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pb6d\" (UniqueName: \"kubernetes.io/projected/cd3f420e-13ec-4e1a-8142-31b0bf64493e-kube-api-access-8pb6d\") pod \"openshift-apiserver-operator-796bbdcf4f-q5b6g\" (UID: \"cd3f420e-13ec-4e1a-8142-31b0bf64493e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599566 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-client-ca\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599724 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-serving-cert\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599852 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd3f420e-13ec-4e1a-8142-31b0bf64493e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-q5b6g\" (UID: \"cd3f420e-13ec-4e1a-8142-31b0bf64493e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599916 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftkq9\" (UniqueName: \"kubernetes.io/projected/cbf444c4-dd3b-4cf7-9771-fea9294124c5-kube-api-access-ftkq9\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.599958 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600182 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/036169eb-913d-470b-b749-6583deb8f396-serving-cert\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600202 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-oauth-serving-cert\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600220 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd3f420e-13ec-4e1a-8142-31b0bf64493e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-q5b6g\" (UID: \"cd3f420e-13ec-4e1a-8142-31b0bf64493e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600339 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-registry-certificates\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600360 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf444c4-dd3b-4cf7-9771-fea9294124c5-serving-cert\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600380 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx2jb\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-kube-api-access-mx2jb\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600397 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-metrics-tls\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600542 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a6ab936d-3cc8-4c09-869a-50783298666d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600558 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-trusted-ca-bundle\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600577 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-etcd-ca\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600690 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9190adfa-bf4f-421b-a2d9-21bb94b94334-serving-cert\") pod \"openshift-config-operator-7777fb866f-skb9s\" (UID: \"9190adfa-bf4f-421b-a2d9-21bb94b94334\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600719 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxjpq\" (UniqueName: \"kubernetes.io/projected/9190adfa-bf4f-421b-a2d9-21bb94b94334-kube-api-access-bxjpq\") pod \"openshift-config-operator-7777fb866f-skb9s\" (UID: \"9190adfa-bf4f-421b-a2d9-21bb94b94334\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600764 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-registry-tls\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600894 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-bound-sa-token\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600928 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp8tg\" (UniqueName: \"kubernetes.io/projected/bf95c26c-10fd-41bb-968a-2b5df4066d29-kube-api-access-cp8tg\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600946 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a20f4d0e-9b06-4bc0-a532-a104392a18dd-serving-cert\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.600965 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a20f4d0e-9b06-4bc0-a532-a104392a18dd-trusted-ca\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:09 crc kubenswrapper[4815]: E1205 09:09:09.603898 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.103873253 +0000 UTC m=+148.982480090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.694240 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.695083 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.695919 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.697674 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.699359 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.704227 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:09 crc kubenswrapper[4815]: E1205 09:09:09.705907 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.205844972 +0000 UTC m=+149.084451809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.706559 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-oauth-config\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.706722 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-etcd-client\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.713716 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.713786 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/59496318-7ad1-4c38-ab56-b51a0ba983dc-srv-cert\") pod \"olm-operator-6b444d44fb-ctjqr\" (UID: \"59496318-7ad1-4c38-ab56-b51a0ba983dc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.713849 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-trusted-ca\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.713880 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.713920 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e9fb212-7595-49b3-8134-4257663737b6-metrics-certs\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.713941 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2rg4\" (UniqueName: \"kubernetes.io/projected/b78558ed-6530-417a-a88e-a17d601593b0-kube-api-access-r2rg4\") pod \"collect-profiles-29415420-98n6w\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.713959 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-config\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714013 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63f1c30-9543-4574-bb51-eb7fb474d495-config\") pod \"service-ca-operator-777779d784-47vsq\" (UID: \"b63f1c30-9543-4574-bb51-eb7fb474d495\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714031 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714120 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714140 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4c34a363-c60d-4d61-a2ea-4d93a54862ba-certs\") pod \"machine-config-server-kj46k\" (UID: \"4c34a363-c60d-4d61-a2ea-4d93a54862ba\") " pod="openshift-machine-config-operator/machine-config-server-kj46k" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714174 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36964d20-077c-4fa4-9672-50b92f34a949-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k6brj\" (UID: \"36964d20-077c-4fa4-9672-50b92f34a949\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714190 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmrrx\" (UniqueName: \"kubernetes.io/projected/3c51b465-c48a-4e6e-b359-7d730bce4980-kube-api-access-mmrrx\") pod \"dns-default-d5gt5\" (UID: \"3c51b465-c48a-4e6e-b359-7d730bce4980\") " pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714208 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c954t\" (UniqueName: \"kubernetes.io/projected/250151de-9f8b-4b4b-af77-b7b35ec0f022-kube-api-access-c954t\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714229 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36964d20-077c-4fa4-9672-50b92f34a949-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k6brj\" (UID: \"36964d20-077c-4fa4-9672-50b92f34a949\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714278 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be45782f-d05a-4f70-8d05-3556a37a9698-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714298 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-client-ca\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714314 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-config\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714332 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714348 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d30f61cc-7d09-4ae2-86c8-7ec8a168dce8-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6vqd5\" (UID: \"d30f61cc-7d09-4ae2-86c8-7ec8a168dce8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714365 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714392 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-client-ca\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714408 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714425 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-serving-cert\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714441 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftkq9\" (UniqueName: \"kubernetes.io/projected/cbf444c4-dd3b-4cf7-9771-fea9294124c5-kube-api-access-ftkq9\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714455 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-dir\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714474 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714503 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2phf\" (UniqueName: \"kubernetes.io/projected/cfe33ba6-894d-4702-b0b9-43abbbe4d6d7-kube-api-access-d2phf\") pod \"ingress-canary-rhtl7\" (UID: \"cfe33ba6-894d-4702-b0b9-43abbbe4d6d7\") " pod="openshift-ingress-canary/ingress-canary-rhtl7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714520 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgwq2\" (UniqueName: \"kubernetes.io/projected/9a9c7304-7002-4b17-9404-d1f17caea8b3-kube-api-access-lgwq2\") pod \"machine-config-controller-84d6567774-zjxb9\" (UID: \"9a9c7304-7002-4b17-9404-d1f17caea8b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714537 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-oauth-serving-cert\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714560 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b63f1c30-9543-4574-bb51-eb7fb474d495-serving-cert\") pod \"service-ca-operator-777779d784-47vsq\" (UID: \"b63f1c30-9543-4574-bb51-eb7fb474d495\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714579 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-registry-certificates\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714595 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf444c4-dd3b-4cf7-9771-fea9294124c5-serving-cert\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714621 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45j9d\" (UniqueName: \"kubernetes.io/projected/4c34a363-c60d-4d61-a2ea-4d93a54862ba-kube-api-access-45j9d\") pod \"machine-config-server-kj46k\" (UID: \"4c34a363-c60d-4d61-a2ea-4d93a54862ba\") " pod="openshift-machine-config-operator/machine-config-server-kj46k" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714637 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpg5k\" (UniqueName: \"kubernetes.io/projected/fcbaef71-6da7-4bf7-9ecf-1438e91571b6-kube-api-access-wpg5k\") pod \"migrator-59844c95c7-tmqvz\" (UID: \"fcbaef71-6da7-4bf7-9ecf-1438e91571b6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714651 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-images\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714668 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a6ab936d-3cc8-4c09-869a-50783298666d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714688 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-trusted-ca-bundle\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714730 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9190adfa-bf4f-421b-a2d9-21bb94b94334-serving-cert\") pod \"openshift-config-operator-7777fb866f-skb9s\" (UID: \"9190adfa-bf4f-421b-a2d9-21bb94b94334\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714749 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxjpq\" (UniqueName: \"kubernetes.io/projected/9190adfa-bf4f-421b-a2d9-21bb94b94334-kube-api-access-bxjpq\") pod \"openshift-config-operator-7777fb866f-skb9s\" (UID: \"9190adfa-bf4f-421b-a2d9-21bb94b94334\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714771 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-bound-sa-token\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714791 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp8tg\" (UniqueName: \"kubernetes.io/projected/bf95c26c-10fd-41bb-968a-2b5df4066d29-kube-api-access-cp8tg\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714812 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f8123ac7-4c54-4dbf-be97-52221d239894-signing-key\") pod \"service-ca-9c57cc56f-4jpx4\" (UID: \"f8123ac7-4c54-4dbf-be97-52221d239894\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714828 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-plugins-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714864 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-proxy-tls\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714883 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6bs4\" (UniqueName: \"kubernetes.io/projected/036169eb-913d-470b-b749-6583deb8f396-kube-api-access-r6bs4\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714933 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-etcd-service-ca\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714951 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/5e9fb212-7595-49b3-8134-4257663737b6-stats-auth\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714969 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-trusted-ca\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.714987 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-policies\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715005 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/90e530b4-0406-491b-8211-f40acd36696b-profile-collector-cert\") pod \"catalog-operator-68c6474976-6q2w4\" (UID: \"90e530b4-0406-491b-8211-f40acd36696b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715022 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8hsj\" (UniqueName: \"kubernetes.io/projected/5e9fb212-7595-49b3-8134-4257663737b6-kube-api-access-j8hsj\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715044 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-bound-sa-token\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715060 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-socket-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715086 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d30f61cc-7d09-4ae2-86c8-7ec8a168dce8-config\") pod \"kube-controller-manager-operator-78b949d7b-6vqd5\" (UID: \"d30f61cc-7d09-4ae2-86c8-7ec8a168dce8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715102 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/43932d3f-4a3d-4ea0-8d0b-1443c876186b-webhook-cert\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715120 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-serving-cert\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715135 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b2cadb7f-14cf-4391-a43a-801a1ade7254-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-szts7\" (UID: \"b2cadb7f-14cf-4391-a43a-801a1ade7254\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715152 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnssk\" (UniqueName: \"kubernetes.io/projected/52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b-kube-api-access-wnssk\") pod \"package-server-manager-789f6589d5-gnsgl\" (UID: \"52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715170 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715202 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715222 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkhcx\" (UniqueName: \"kubernetes.io/projected/b63f1c30-9543-4574-bb51-eb7fb474d495-kube-api-access-kkhcx\") pod \"service-ca-operator-777779d784-47vsq\" (UID: \"b63f1c30-9543-4574-bb51-eb7fb474d495\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715239 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jj8r\" (UniqueName: \"kubernetes.io/projected/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-kube-api-access-4jj8r\") pod \"marketplace-operator-79b997595-dgk4c\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715255 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq29x\" (UniqueName: \"kubernetes.io/projected/90e530b4-0406-491b-8211-f40acd36696b-kube-api-access-dq29x\") pod \"catalog-operator-68c6474976-6q2w4\" (UID: \"90e530b4-0406-491b-8211-f40acd36696b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715269 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gnsgl\" (UID: \"52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715284 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b78558ed-6530-417a-a88e-a17d601593b0-secret-volume\") pod \"collect-profiles-29415420-98n6w\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715314 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2211eb63-8b44-48c6-9cca-c032127ddb95-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kvstg\" (UID: \"2211eb63-8b44-48c6-9cca-c032127ddb95\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715330 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfccw\" (UniqueName: \"kubernetes.io/projected/025c3afb-3b9a-4fe8-a11c-50b309374464-kube-api-access-gfccw\") pod \"kube-storage-version-migrator-operator-b67b599dd-8vfp6\" (UID: \"025c3afb-3b9a-4fe8-a11c-50b309374464\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715346 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3c51b465-c48a-4e6e-b359-7d730bce4980-metrics-tls\") pod \"dns-default-d5gt5\" (UID: \"3c51b465-c48a-4e6e-b359-7d730bce4980\") " pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715361 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/5e9fb212-7595-49b3-8134-4257663737b6-default-certificate\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715378 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5p2pc\" (UniqueName: \"kubernetes.io/projected/59496318-7ad1-4c38-ab56-b51a0ba983dc-kube-api-access-5p2pc\") pod \"olm-operator-6b444d44fb-ctjqr\" (UID: \"59496318-7ad1-4c38-ab56-b51a0ba983dc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715398 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-csi-data-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715412 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b78558ed-6530-417a-a88e-a17d601593b0-config-volume\") pod \"collect-profiles-29415420-98n6w\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715430 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a6ab936d-3cc8-4c09-869a-50783298666d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715448 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2cadb7f-14cf-4391-a43a-801a1ade7254-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-szts7\" (UID: \"b2cadb7f-14cf-4391-a43a-801a1ade7254\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715462 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/59496318-7ad1-4c38-ab56-b51a0ba983dc-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ctjqr\" (UID: \"59496318-7ad1-4c38-ab56-b51a0ba983dc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715480 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrd58\" (UniqueName: \"kubernetes.io/projected/2211eb63-8b44-48c6-9cca-c032127ddb95-kube-api-access-hrd58\") pod \"openshift-controller-manager-operator-756b6f6bc6-kvstg\" (UID: \"2211eb63-8b44-48c6-9cca-c032127ddb95\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715510 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/be45782f-d05a-4f70-8d05-3556a37a9698-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715526 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025c3afb-3b9a-4fe8-a11c-50b309374464-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8vfp6\" (UID: \"025c3afb-3b9a-4fe8-a11c-50b309374464\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715543 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/be45782f-d05a-4f70-8d05-3556a37a9698-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715558 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715573 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2cadb7f-14cf-4391-a43a-801a1ade7254-config\") pod \"kube-apiserver-operator-766d6c64bb-szts7\" (UID: \"b2cadb7f-14cf-4391-a43a-801a1ade7254\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715589 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2znbq\" (UniqueName: \"kubernetes.io/projected/be45782f-d05a-4f70-8d05-3556a37a9698-kube-api-access-2znbq\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715607 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s4ljm\" (UniqueName: \"kubernetes.io/projected/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-kube-api-access-s4ljm\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715623 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/90e530b4-0406-491b-8211-f40acd36696b-srv-cert\") pod \"catalog-operator-68c6474976-6q2w4\" (UID: \"90e530b4-0406-491b-8211-f40acd36696b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715641 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c51b465-c48a-4e6e-b359-7d730bce4980-config-volume\") pod \"dns-default-d5gt5\" (UID: \"3c51b465-c48a-4e6e-b359-7d730bce4980\") " pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715656 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e9fb212-7595-49b3-8134-4257663737b6-service-ca-bundle\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715673 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9190adfa-bf4f-421b-a2d9-21bb94b94334-available-featuregates\") pod \"openshift-config-operator-7777fb866f-skb9s\" (UID: \"9190adfa-bf4f-421b-a2d9-21bb94b94334\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715687 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-config\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715708 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cf7wz\" (UniqueName: \"kubernetes.io/projected/a20f4d0e-9b06-4bc0-a532-a104392a18dd-kube-api-access-cf7wz\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715728 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pb6d\" (UniqueName: \"kubernetes.io/projected/cd3f420e-13ec-4e1a-8142-31b0bf64493e-kube-api-access-8pb6d\") pod \"openshift-apiserver-operator-796bbdcf4f-q5b6g\" (UID: \"cd3f420e-13ec-4e1a-8142-31b0bf64493e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715747 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4c34a363-c60d-4d61-a2ea-4d93a54862ba-node-bootstrap-token\") pod \"machine-config-server-kj46k\" (UID: \"4c34a363-c60d-4d61-a2ea-4d93a54862ba\") " pod="openshift-machine-config-operator/machine-config-server-kj46k" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715768 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-registration-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715786 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9a9c7304-7002-4b17-9404-d1f17caea8b3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zjxb9\" (UID: \"9a9c7304-7002-4b17-9404-d1f17caea8b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715807 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/43932d3f-4a3d-4ea0-8d0b-1443c876186b-tmpfs\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715829 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd3f420e-13ec-4e1a-8142-31b0bf64493e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-q5b6g\" (UID: \"cd3f420e-13ec-4e1a-8142-31b0bf64493e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715858 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715875 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715891 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36964d20-077c-4fa4-9672-50b92f34a949-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k6brj\" (UID: \"36964d20-077c-4fa4-9672-50b92f34a949\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715907 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/036169eb-913d-470b-b749-6583deb8f396-serving-cert\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715923 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd3f420e-13ec-4e1a-8142-31b0bf64493e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-q5b6g\" (UID: \"cd3f420e-13ec-4e1a-8142-31b0bf64493e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715938 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715955 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2427\" (UniqueName: \"kubernetes.io/projected/a4d093b9-c855-48b0-9d44-c43fe89793bb-kube-api-access-x2427\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715972 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx2jb\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-kube-api-access-mx2jb\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.715988 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-metrics-tls\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716004 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-mountpoint-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716022 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-etcd-ca\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716038 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/43932d3f-4a3d-4ea0-8d0b-1443c876186b-apiservice-cert\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716072 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-registry-tls\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716087 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d30f61cc-7d09-4ae2-86c8-7ec8a168dce8-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6vqd5\" (UID: \"d30f61cc-7d09-4ae2-86c8-7ec8a168dce8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716103 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a20f4d0e-9b06-4bc0-a532-a104392a18dd-serving-cert\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716119 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a20f4d0e-9b06-4bc0-a532-a104392a18dd-trusted-ca\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716136 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ec745360-4a3f-4af0-ab12-588ef1345804-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-cgjsg\" (UID: \"ec745360-4a3f-4af0-ab12-588ef1345804\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716150 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cfe33ba6-894d-4702-b0b9-43abbbe4d6d7-cert\") pod \"ingress-canary-rhtl7\" (UID: \"cfe33ba6-894d-4702-b0b9-43abbbe4d6d7\") " pod="openshift-ingress-canary/ingress-canary-rhtl7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716176 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dgk4c\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716191 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9a9c7304-7002-4b17-9404-d1f17caea8b3-proxy-tls\") pod \"machine-config-controller-84d6567774-zjxb9\" (UID: \"9a9c7304-7002-4b17-9404-d1f17caea8b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716206 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbcn4\" (UniqueName: \"kubernetes.io/projected/43932d3f-4a3d-4ea0-8d0b-1443c876186b-kube-api-access-bbcn4\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716224 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb48516f-711a-4131-a555-de5ee6e96ae1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-h89pl\" (UID: \"fb48516f-711a-4131-a555-de5ee6e96ae1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716246 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-service-ca\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716262 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-config\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716279 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6v6w\" (UniqueName: \"kubernetes.io/projected/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-kube-api-access-d6v6w\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716296 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85tbf\" (UniqueName: \"kubernetes.io/projected/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-kube-api-access-85tbf\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716313 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dgk4c\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716330 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a20f4d0e-9b06-4bc0-a532-a104392a18dd-config\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716346 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2211eb63-8b44-48c6-9cca-c032127ddb95-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kvstg\" (UID: \"2211eb63-8b44-48c6-9cca-c032127ddb95\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716362 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f8123ac7-4c54-4dbf-be97-52221d239894-signing-cabundle\") pod \"service-ca-9c57cc56f-4jpx4\" (UID: \"f8123ac7-4c54-4dbf-be97-52221d239894\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716378 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr7q5\" (UniqueName: \"kubernetes.io/projected/ec745360-4a3f-4af0-ab12-588ef1345804-kube-api-access-tr7q5\") pod \"multus-admission-controller-857f4d67dd-cgjsg\" (UID: \"ec745360-4a3f-4af0-ab12-588ef1345804\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716394 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ksr7q\" (UniqueName: \"kubernetes.io/projected/f8123ac7-4c54-4dbf-be97-52221d239894-kube-api-access-ksr7q\") pod \"service-ca-9c57cc56f-4jpx4\" (UID: \"f8123ac7-4c54-4dbf-be97-52221d239894\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716413 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025c3afb-3b9a-4fe8-a11c-50b309374464-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8vfp6\" (UID: \"025c3afb-3b9a-4fe8-a11c-50b309374464\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.716430 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhdxx\" (UniqueName: \"kubernetes.io/projected/fb48516f-711a-4131-a555-de5ee6e96ae1-kube-api-access-dhdxx\") pod \"control-plane-machine-set-operator-78cbb6b69f-h89pl\" (UID: \"fb48516f-711a-4131-a555-de5ee6e96ae1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.710511 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-etcd-client\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.710889 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-oauth-config\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.717603 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-trusted-ca\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.718903 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-config\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.719985 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vd7n\" (UniqueName: \"kubernetes.io/projected/225d33d1-7da4-4937-a843-7594d7549454-kube-api-access-4vd7n\") pod \"dns-operator-744455d44c-ltjzv\" (UID: \"225d33d1-7da4-4937-a843-7594d7549454\") " pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.719982 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a6ab936d-3cc8-4c09-869a-50783298666d-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.720359 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/9190adfa-bf4f-421b-a2d9-21bb94b94334-available-featuregates\") pod \"openshift-config-operator-7777fb866f-skb9s\" (UID: \"9190adfa-bf4f-421b-a2d9-21bb94b94334\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.720881 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-etcd-ca\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.721311 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-etcd-service-ca\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.721820 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a20f4d0e-9b06-4bc0-a532-a104392a18dd-trusted-ca\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.722679 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-client-ca\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.722980 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-client-ca\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.723460 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/036169eb-913d-470b-b749-6583deb8f396-serving-cert\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.724428 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-config\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.725568 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-config\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.726457 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a20f4d0e-9b06-4bc0-a532-a104392a18dd-config\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.726966 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-trusted-ca\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.727529 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/be45782f-d05a-4f70-8d05-3556a37a9698-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.728257 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd3f420e-13ec-4e1a-8142-31b0bf64493e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-q5b6g\" (UID: \"cd3f420e-13ec-4e1a-8142-31b0bf64493e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.729084 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd3f420e-13ec-4e1a-8142-31b0bf64493e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-q5b6g\" (UID: \"cd3f420e-13ec-4e1a-8142-31b0bf64493e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.729791 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-config\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: E1205 09:09:09.730293 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.230274481 +0000 UTC m=+149.108881318 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.730616 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a20f4d0e-9b06-4bc0-a532-a104392a18dd-serving-cert\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.731380 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-registry-tls\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.732831 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-service-ca\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.734388 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2211eb63-8b44-48c6-9cca-c032127ddb95-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-kvstg\" (UID: \"2211eb63-8b44-48c6-9cca-c032127ddb95\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.735134 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-oauth-serving-cert\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.735889 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.736000 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9190adfa-bf4f-421b-a2d9-21bb94b94334-serving-cert\") pod \"openshift-config-operator-7777fb866f-skb9s\" (UID: \"9190adfa-bf4f-421b-a2d9-21bb94b94334\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.736886 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2211eb63-8b44-48c6-9cca-c032127ddb95-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-kvstg\" (UID: \"2211eb63-8b44-48c6-9cca-c032127ddb95\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.737037 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-trusted-ca-bundle\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.738250 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-registry-certificates\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.738807 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a6ab936d-3cc8-4c09-869a-50783298666d-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.738922 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf444c4-dd3b-4cf7-9771-fea9294124c5-serving-cert\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.739066 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be45782f-d05a-4f70-8d05-3556a37a9698-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.754367 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-serving-cert\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.760186 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrxtr\" (UniqueName: \"kubernetes.io/projected/e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4-kube-api-access-wrxtr\") pod \"authentication-operator-69f744f599-fvz7z\" (UID: \"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.765914 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-metrics-tls\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.768644 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2q5t\" (UniqueName: \"kubernetes.io/projected/97c971b0-67a2-456b-ae6b-eb4f2a69630d-kube-api-access-n2q5t\") pod \"cluster-samples-operator-665b6dd947-2w9rk\" (UID: \"97c971b0-67a2-456b-ae6b-eb4f2a69630d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.769942 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-serving-cert\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.782158 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.783423 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.788069 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vsrn\" (UniqueName: \"kubernetes.io/projected/1615b53d-0e1b-4d6a-a369-1a614711721c-kube-api-access-2vsrn\") pod \"machine-approver-56656f9798-tnfbl\" (UID: \"1615b53d-0e1b-4d6a-a369-1a614711721c\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.803465 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.817692 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.817865 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63f1c30-9543-4574-bb51-eb7fb474d495-config\") pod \"service-ca-operator-777779d784-47vsq\" (UID: \"b63f1c30-9543-4574-bb51-eb7fb474d495\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.817884 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4c34a363-c60d-4d61-a2ea-4d93a54862ba-certs\") pod \"machine-config-server-kj46k\" (UID: \"4c34a363-c60d-4d61-a2ea-4d93a54862ba\") " pod="openshift-machine-config-operator/machine-config-server-kj46k" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.817907 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36964d20-077c-4fa4-9672-50b92f34a949-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k6brj\" (UID: \"36964d20-077c-4fa4-9672-50b92f34a949\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.817922 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmrrx\" (UniqueName: \"kubernetes.io/projected/3c51b465-c48a-4e6e-b359-7d730bce4980-kube-api-access-mmrrx\") pod \"dns-default-d5gt5\" (UID: \"3c51b465-c48a-4e6e-b359-7d730bce4980\") " pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.817941 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c954t\" (UniqueName: \"kubernetes.io/projected/250151de-9f8b-4b4b-af77-b7b35ec0f022-kube-api-access-c954t\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.817963 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.817987 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818008 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36964d20-077c-4fa4-9672-50b92f34a949-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k6brj\" (UID: \"36964d20-077c-4fa4-9672-50b92f34a949\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818036 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818056 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d30f61cc-7d09-4ae2-86c8-7ec8a168dce8-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6vqd5\" (UID: \"d30f61cc-7d09-4ae2-86c8-7ec8a168dce8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818079 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818107 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818126 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-dir\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818141 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818156 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2phf\" (UniqueName: \"kubernetes.io/projected/cfe33ba6-894d-4702-b0b9-43abbbe4d6d7-kube-api-access-d2phf\") pod \"ingress-canary-rhtl7\" (UID: \"cfe33ba6-894d-4702-b0b9-43abbbe4d6d7\") " pod="openshift-ingress-canary/ingress-canary-rhtl7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818171 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgwq2\" (UniqueName: \"kubernetes.io/projected/9a9c7304-7002-4b17-9404-d1f17caea8b3-kube-api-access-lgwq2\") pod \"machine-config-controller-84d6567774-zjxb9\" (UID: \"9a9c7304-7002-4b17-9404-d1f17caea8b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818187 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b63f1c30-9543-4574-bb51-eb7fb474d495-serving-cert\") pod \"service-ca-operator-777779d784-47vsq\" (UID: \"b63f1c30-9543-4574-bb51-eb7fb474d495\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818205 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45j9d\" (UniqueName: \"kubernetes.io/projected/4c34a363-c60d-4d61-a2ea-4d93a54862ba-kube-api-access-45j9d\") pod \"machine-config-server-kj46k\" (UID: \"4c34a363-c60d-4d61-a2ea-4d93a54862ba\") " pod="openshift-machine-config-operator/machine-config-server-kj46k" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818221 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpg5k\" (UniqueName: \"kubernetes.io/projected/fcbaef71-6da7-4bf7-9ecf-1438e91571b6-kube-api-access-wpg5k\") pod \"migrator-59844c95c7-tmqvz\" (UID: \"fcbaef71-6da7-4bf7-9ecf-1438e91571b6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818237 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-images\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818268 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f8123ac7-4c54-4dbf-be97-52221d239894-signing-key\") pod \"service-ca-9c57cc56f-4jpx4\" (UID: \"f8123ac7-4c54-4dbf-be97-52221d239894\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818281 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-plugins-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818295 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-proxy-tls\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818316 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/5e9fb212-7595-49b3-8134-4257663737b6-stats-auth\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818331 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8hsj\" (UniqueName: \"kubernetes.io/projected/5e9fb212-7595-49b3-8134-4257663737b6-kube-api-access-j8hsj\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818346 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-policies\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818360 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/90e530b4-0406-491b-8211-f40acd36696b-profile-collector-cert\") pod \"catalog-operator-68c6474976-6q2w4\" (UID: \"90e530b4-0406-491b-8211-f40acd36696b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818381 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-socket-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818395 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d30f61cc-7d09-4ae2-86c8-7ec8a168dce8-config\") pod \"kube-controller-manager-operator-78b949d7b-6vqd5\" (UID: \"d30f61cc-7d09-4ae2-86c8-7ec8a168dce8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818412 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/43932d3f-4a3d-4ea0-8d0b-1443c876186b-webhook-cert\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818428 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnssk\" (UniqueName: \"kubernetes.io/projected/52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b-kube-api-access-wnssk\") pod \"package-server-manager-789f6589d5-gnsgl\" (UID: \"52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818442 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b2cadb7f-14cf-4391-a43a-801a1ade7254-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-szts7\" (UID: \"b2cadb7f-14cf-4391-a43a-801a1ade7254\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818456 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818470 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkhcx\" (UniqueName: \"kubernetes.io/projected/b63f1c30-9543-4574-bb51-eb7fb474d495-kube-api-access-kkhcx\") pod \"service-ca-operator-777779d784-47vsq\" (UID: \"b63f1c30-9543-4574-bb51-eb7fb474d495\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818491 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jj8r\" (UniqueName: \"kubernetes.io/projected/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-kube-api-access-4jj8r\") pod \"marketplace-operator-79b997595-dgk4c\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818520 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b78558ed-6530-417a-a88e-a17d601593b0-secret-volume\") pod \"collect-profiles-29415420-98n6w\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818547 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq29x\" (UniqueName: \"kubernetes.io/projected/90e530b4-0406-491b-8211-f40acd36696b-kube-api-access-dq29x\") pod \"catalog-operator-68c6474976-6q2w4\" (UID: \"90e530b4-0406-491b-8211-f40acd36696b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818561 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gnsgl\" (UID: \"52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818581 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfccw\" (UniqueName: \"kubernetes.io/projected/025c3afb-3b9a-4fe8-a11c-50b309374464-kube-api-access-gfccw\") pod \"kube-storage-version-migrator-operator-b67b599dd-8vfp6\" (UID: \"025c3afb-3b9a-4fe8-a11c-50b309374464\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818594 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3c51b465-c48a-4e6e-b359-7d730bce4980-metrics-tls\") pod \"dns-default-d5gt5\" (UID: \"3c51b465-c48a-4e6e-b359-7d730bce4980\") " pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818609 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/5e9fb212-7595-49b3-8134-4257663737b6-default-certificate\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818626 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5p2pc\" (UniqueName: \"kubernetes.io/projected/59496318-7ad1-4c38-ab56-b51a0ba983dc-kube-api-access-5p2pc\") pod \"olm-operator-6b444d44fb-ctjqr\" (UID: \"59496318-7ad1-4c38-ab56-b51a0ba983dc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818639 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-csi-data-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818653 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b78558ed-6530-417a-a88e-a17d601593b0-config-volume\") pod \"collect-profiles-29415420-98n6w\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818683 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/59496318-7ad1-4c38-ab56-b51a0ba983dc-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ctjqr\" (UID: \"59496318-7ad1-4c38-ab56-b51a0ba983dc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818697 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2cadb7f-14cf-4391-a43a-801a1ade7254-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-szts7\" (UID: \"b2cadb7f-14cf-4391-a43a-801a1ade7254\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818714 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025c3afb-3b9a-4fe8-a11c-50b309374464-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8vfp6\" (UID: \"025c3afb-3b9a-4fe8-a11c-50b309374464\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818738 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818753 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2cadb7f-14cf-4391-a43a-801a1ade7254-config\") pod \"kube-apiserver-operator-766d6c64bb-szts7\" (UID: \"b2cadb7f-14cf-4391-a43a-801a1ade7254\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818772 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s4ljm\" (UniqueName: \"kubernetes.io/projected/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-kube-api-access-s4ljm\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818786 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/90e530b4-0406-491b-8211-f40acd36696b-srv-cert\") pod \"catalog-operator-68c6474976-6q2w4\" (UID: \"90e530b4-0406-491b-8211-f40acd36696b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.818800 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c51b465-c48a-4e6e-b359-7d730bce4980-config-volume\") pod \"dns-default-d5gt5\" (UID: \"3c51b465-c48a-4e6e-b359-7d730bce4980\") " pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:09 crc kubenswrapper[4815]: E1205 09:09:09.818911 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.318896272 +0000 UTC m=+149.197503119 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.819669 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e9fb212-7595-49b3-8134-4257663737b6-service-ca-bundle\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.819749 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4c34a363-c60d-4d61-a2ea-4d93a54862ba-node-bootstrap-token\") pod \"machine-config-server-kj46k\" (UID: \"4c34a363-c60d-4d61-a2ea-4d93a54862ba\") " pod="openshift-machine-config-operator/machine-config-server-kj46k" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.819798 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/43932d3f-4a3d-4ea0-8d0b-1443c876186b-tmpfs\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.819829 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-registration-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.819878 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9a9c7304-7002-4b17-9404-d1f17caea8b3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zjxb9\" (UID: \"9a9c7304-7002-4b17-9404-d1f17caea8b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.819925 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.819953 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36964d20-077c-4fa4-9672-50b92f34a949-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k6brj\" (UID: \"36964d20-077c-4fa4-9672-50b92f34a949\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.819988 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820017 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820046 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2427\" (UniqueName: \"kubernetes.io/projected/a4d093b9-c855-48b0-9d44-c43fe89793bb-kube-api-access-x2427\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820074 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-mountpoint-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820097 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/43932d3f-4a3d-4ea0-8d0b-1443c876186b-apiservice-cert\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820121 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d30f61cc-7d09-4ae2-86c8-7ec8a168dce8-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6vqd5\" (UID: \"d30f61cc-7d09-4ae2-86c8-7ec8a168dce8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820157 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ec745360-4a3f-4af0-ab12-588ef1345804-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-cgjsg\" (UID: \"ec745360-4a3f-4af0-ab12-588ef1345804\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820180 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cfe33ba6-894d-4702-b0b9-43abbbe4d6d7-cert\") pod \"ingress-canary-rhtl7\" (UID: \"cfe33ba6-894d-4702-b0b9-43abbbe4d6d7\") " pod="openshift-ingress-canary/ingress-canary-rhtl7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820205 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dgk4c\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820228 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9a9c7304-7002-4b17-9404-d1f17caea8b3-proxy-tls\") pod \"machine-config-controller-84d6567774-zjxb9\" (UID: \"9a9c7304-7002-4b17-9404-d1f17caea8b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820440 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbcn4\" (UniqueName: \"kubernetes.io/projected/43932d3f-4a3d-4ea0-8d0b-1443c876186b-kube-api-access-bbcn4\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820552 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb48516f-711a-4131-a555-de5ee6e96ae1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-h89pl\" (UID: \"fb48516f-711a-4131-a555-de5ee6e96ae1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820634 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dgk4c\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820673 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f8123ac7-4c54-4dbf-be97-52221d239894-signing-cabundle\") pod \"service-ca-9c57cc56f-4jpx4\" (UID: \"f8123ac7-4c54-4dbf-be97-52221d239894\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820700 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ksr7q\" (UniqueName: \"kubernetes.io/projected/f8123ac7-4c54-4dbf-be97-52221d239894-kube-api-access-ksr7q\") pod \"service-ca-9c57cc56f-4jpx4\" (UID: \"f8123ac7-4c54-4dbf-be97-52221d239894\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820723 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025c3afb-3b9a-4fe8-a11c-50b309374464-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8vfp6\" (UID: \"025c3afb-3b9a-4fe8-a11c-50b309374464\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820750 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr7q5\" (UniqueName: \"kubernetes.io/projected/ec745360-4a3f-4af0-ab12-588ef1345804-kube-api-access-tr7q5\") pod \"multus-admission-controller-857f4d67dd-cgjsg\" (UID: \"ec745360-4a3f-4af0-ab12-588ef1345804\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820775 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhdxx\" (UniqueName: \"kubernetes.io/projected/fb48516f-711a-4131-a555-de5ee6e96ae1-kube-api-access-dhdxx\") pod \"control-plane-machine-set-operator-78cbb6b69f-h89pl\" (UID: \"fb48516f-711a-4131-a555-de5ee6e96ae1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820802 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820804 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-auth-proxy-config\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820825 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/59496318-7ad1-4c38-ab56-b51a0ba983dc-srv-cert\") pod \"olm-operator-6b444d44fb-ctjqr\" (UID: \"59496318-7ad1-4c38-ab56-b51a0ba983dc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820857 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e9fb212-7595-49b3-8134-4257663737b6-metrics-certs\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820881 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2rg4\" (UniqueName: \"kubernetes.io/projected/b78558ed-6530-417a-a88e-a17d601593b0-kube-api-access-r2rg4\") pod \"collect-profiles-29415420-98n6w\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.820909 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.824451 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/ec745360-4a3f-4af0-ab12-588ef1345804-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-cgjsg\" (UID: \"ec745360-4a3f-4af0-ab12-588ef1345804\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.824478 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-socket-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.825042 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d30f61cc-7d09-4ae2-86c8-7ec8a168dce8-config\") pod \"kube-controller-manager-operator-78b949d7b-6vqd5\" (UID: \"d30f61cc-7d09-4ae2-86c8-7ec8a168dce8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.825375 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-dgk4c\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.819477 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b63f1c30-9543-4574-bb51-eb7fb474d495-config\") pod \"service-ca-operator-777779d784-47vsq\" (UID: \"b63f1c30-9543-4574-bb51-eb7fb474d495\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.826102 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/90e530b4-0406-491b-8211-f40acd36696b-profile-collector-cert\") pod \"catalog-operator-68c6474976-6q2w4\" (UID: \"90e530b4-0406-491b-8211-f40acd36696b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.827365 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.827769 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9a9c7304-7002-4b17-9404-d1f17caea8b3-proxy-tls\") pod \"machine-config-controller-84d6567774-zjxb9\" (UID: \"9a9c7304-7002-4b17-9404-d1f17caea8b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.827784 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/43932d3f-4a3d-4ea0-8d0b-1443c876186b-webhook-cert\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.828092 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e9fb212-7595-49b3-8134-4257663737b6-service-ca-bundle\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.828125 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.841401 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-dir\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.849096 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b63f1c30-9543-4574-bb51-eb7fb474d495-serving-cert\") pod \"service-ca-operator-777779d784-47vsq\" (UID: \"b63f1c30-9543-4574-bb51-eb7fb474d495\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.849650 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.849850 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-dgk4c\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.850513 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f8123ac7-4c54-4dbf-be97-52221d239894-signing-cabundle\") pod \"service-ca-9c57cc56f-4jpx4\" (UID: \"f8123ac7-4c54-4dbf-be97-52221d239894\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.852911 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.853590 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-images\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.853802 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.903372 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36964d20-077c-4fa4-9672-50b92f34a949-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k6brj\" (UID: \"36964d20-077c-4fa4-9672-50b92f34a949\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.903894 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/4c34a363-c60d-4d61-a2ea-4d93a54862ba-certs\") pod \"machine-config-server-kj46k\" (UID: \"4c34a363-c60d-4d61-a2ea-4d93a54862ba\") " pod="openshift-machine-config-operator/machine-config-server-kj46k" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.904803 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/43932d3f-4a3d-4ea0-8d0b-1443c876186b-tmpfs\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.904863 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-registration-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.905420 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9a9c7304-7002-4b17-9404-d1f17caea8b3-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zjxb9\" (UID: \"9a9c7304-7002-4b17-9404-d1f17caea8b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.914171 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.915317 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36964d20-077c-4fa4-9672-50b92f34a949-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k6brj\" (UID: \"36964d20-077c-4fa4-9672-50b92f34a949\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" Dec 05 09:09:09 crc kubenswrapper[4815]: E1205 09:09:09.915561 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.415547434 +0000 UTC m=+149.294154271 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.918846 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-mountpoint-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.919753 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5e9fb212-7595-49b3-8134-4257663737b6-metrics-certs\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.922023 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-csi-data-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.923072 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gnsgl\" (UID: \"52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.923107 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b78558ed-6530-417a-a88e-a17d601593b0-config-volume\") pod \"collect-profiles-29415420-98n6w\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.924065 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c51b465-c48a-4e6e-b359-7d730bce4980-config-volume\") pod \"dns-default-d5gt5\" (UID: \"3c51b465-c48a-4e6e-b359-7d730bce4980\") " pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.924467 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/43932d3f-4a3d-4ea0-8d0b-1443c876186b-apiservice-cert\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.925183 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/4c34a363-c60d-4d61-a2ea-4d93a54862ba-node-bootstrap-token\") pod \"machine-config-server-kj46k\" (UID: \"4c34a363-c60d-4d61-a2ea-4d93a54862ba\") " pod="openshift-machine-config-operator/machine-config-server-kj46k" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.925770 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/59496318-7ad1-4c38-ab56-b51a0ba983dc-profile-collector-cert\") pod \"olm-operator-6b444d44fb-ctjqr\" (UID: \"59496318-7ad1-4c38-ab56-b51a0ba983dc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.927265 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d30f61cc-7d09-4ae2-86c8-7ec8a168dce8-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-6vqd5\" (UID: \"d30f61cc-7d09-4ae2-86c8-7ec8a168dce8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.927682 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/5e9fb212-7595-49b3-8134-4257663737b6-stats-auth\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.928014 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.928744 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025c3afb-3b9a-4fe8-a11c-50b309374464-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-8vfp6\" (UID: \"025c3afb-3b9a-4fe8-a11c-50b309374464\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.930422 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.930562 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/250151de-9f8b-4b4b-af77-b7b35ec0f022-plugins-dir\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.930758 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.931320 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.934273 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.930910 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.936068 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2cadb7f-14cf-4391-a43a-801a1ade7254-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-szts7\" (UID: \"b2cadb7f-14cf-4391-a43a-801a1ade7254\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.936678 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-proxy-tls\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.930947 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 05 09:09:09 crc kubenswrapper[4815]: E1205 09:09:09.938865 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.438847388 +0000 UTC m=+149.317454225 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.939308 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/90e530b4-0406-491b-8211-f40acd36696b-srv-cert\") pod \"catalog-operator-68c6474976-6q2w4\" (UID: \"90e530b4-0406-491b-8211-f40acd36696b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.939359 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2cadb7f-14cf-4391-a43a-801a1ade7254-config\") pod \"kube-apiserver-operator-766d6c64bb-szts7\" (UID: \"b2cadb7f-14cf-4391-a43a-801a1ade7254\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.939744 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:09 crc kubenswrapper[4815]: E1205 09:09:09.940047 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.440038664 +0000 UTC m=+149.318645501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.944816 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3c51b465-c48a-4e6e-b359-7d730bce4980-metrics-tls\") pod \"dns-default-d5gt5\" (UID: \"3c51b465-c48a-4e6e-b359-7d730bce4980\") " pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.947186 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f8123ac7-4c54-4dbf-be97-52221d239894-signing-key\") pod \"service-ca-9c57cc56f-4jpx4\" (UID: \"f8123ac7-4c54-4dbf-be97-52221d239894\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.947554 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/59496318-7ad1-4c38-ab56-b51a0ba983dc-srv-cert\") pod \"olm-operator-6b444d44fb-ctjqr\" (UID: \"59496318-7ad1-4c38-ab56-b51a0ba983dc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.947807 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b78558ed-6530-417a-a88e-a17d601593b0-secret-volume\") pod \"collect-profiles-29415420-98n6w\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.956628 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.958695 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.958846 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.962854 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.965560 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/fb48516f-711a-4131-a555-de5ee6e96ae1-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-h89pl\" (UID: \"fb48516f-711a-4131-a555-de5ee6e96ae1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.965960 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/5e9fb212-7595-49b3-8134-4257663737b6-default-certificate\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.966263 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.965065 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.974661 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.986985 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025c3afb-3b9a-4fe8-a11c-50b309374464-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-8vfp6\" (UID: \"025c3afb-3b9a-4fe8-a11c-50b309374464\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.988616 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.988749 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.989895 4815 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 05 09:09:09 crc kubenswrapper[4815]: I1205 09:09:09.990711 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-policies\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.009893 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.026548 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.027724 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.056997 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:10 crc kubenswrapper[4815]: E1205 09:09:10.058196 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.558173386 +0000 UTC m=+149.436780213 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.060715 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:10 crc kubenswrapper[4815]: E1205 09:09:10.061115 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.561100763 +0000 UTC m=+149.439707600 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.074122 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/cfe33ba6-894d-4702-b0b9-43abbbe4d6d7-cert\") pod \"ingress-canary-rhtl7\" (UID: \"cfe33ba6-894d-4702-b0b9-43abbbe4d6d7\") " pod="openshift-ingress-canary/ingress-canary-rhtl7" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.092228 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2znbq\" (UniqueName: \"kubernetes.io/projected/be45782f-d05a-4f70-8d05-3556a37a9698-kube-api-access-2znbq\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.109451 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6bs4\" (UniqueName: \"kubernetes.io/projected/036169eb-913d-470b-b749-6583deb8f396-kube-api-access-r6bs4\") pod \"controller-manager-879f6c89f-h66pq\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.119970 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-bound-sa-token\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.150817 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6v6w\" (UniqueName: \"kubernetes.io/projected/a0119b9d-7402-4360-98fa-e5c2b1ecf9ec-kube-api-access-d6v6w\") pod \"etcd-operator-b45778765-wjxhv\" (UID: \"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.166416 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:10 crc kubenswrapper[4815]: E1205 09:09:10.167055 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.667035741 +0000 UTC m=+149.545642588 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.184934 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85tbf\" (UniqueName: \"kubernetes.io/projected/f2e6f1de-e80e-45b3-8caa-ca2c16a027bc-kube-api-access-85tbf\") pod \"ingress-operator-5b745b69d9-j8nhp\" (UID: \"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.191699 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pb6d\" (UniqueName: \"kubernetes.io/projected/cd3f420e-13ec-4e1a-8142-31b0bf64493e-kube-api-access-8pb6d\") pod \"openshift-apiserver-operator-796bbdcf4f-q5b6g\" (UID: \"cd3f420e-13ec-4e1a-8142-31b0bf64493e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.220997 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.228047 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cf7wz\" (UniqueName: \"kubernetes.io/projected/a20f4d0e-9b06-4bc0-a532-a104392a18dd-kube-api-access-cf7wz\") pod \"console-operator-58897d9998-4mvlm\" (UID: \"a20f4d0e-9b06-4bc0-a532-a104392a18dd\") " pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.249003 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/be45782f-d05a-4f70-8d05-3556a37a9698-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-hxnqx\" (UID: \"be45782f-d05a-4f70-8d05-3556a37a9698\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.249235 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.258649 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrd58\" (UniqueName: \"kubernetes.io/projected/2211eb63-8b44-48c6-9cca-c032127ddb95-kube-api-access-hrd58\") pod \"openshift-controller-manager-operator-756b6f6bc6-kvstg\" (UID: \"2211eb63-8b44-48c6-9cca-c032127ddb95\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.258835 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.263566 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.270040 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:10 crc kubenswrapper[4815]: E1205 09:09:10.270411 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.770399491 +0000 UTC m=+149.649006328 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.270749 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx2jb\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-kube-api-access-mx2jb\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.270779 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.288742 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp8tg\" (UniqueName: \"kubernetes.io/projected/bf95c26c-10fd-41bb-968a-2b5df4066d29-kube-api-access-cp8tg\") pod \"console-f9d7485db-d2j7q\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.307740 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.330654 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-bound-sa-token\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.369200 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxjpq\" (UniqueName: \"kubernetes.io/projected/9190adfa-bf4f-421b-a2d9-21bb94b94334-kube-api-access-bxjpq\") pod \"openshift-config-operator-7777fb866f-skb9s\" (UID: \"9190adfa-bf4f-421b-a2d9-21bb94b94334\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.371809 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:10 crc kubenswrapper[4815]: E1205 09:09:10.372193 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:10.872176256 +0000 UTC m=+149.750783093 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.376304 4815 generic.go:334] "Generic (PLEG): container finished" podID="99bb7807-d522-499e-863e-7be3e10ac6ed" containerID="ec1bb1a5ab524a8d993305ae562f2a9f279e43c1b1089d37c70d9f6032e11948" exitCode=0 Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.382617 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" event={"ID":"99bb7807-d522-499e-863e-7be3e10ac6ed","Type":"ContainerDied","Data":"ec1bb1a5ab524a8d993305ae562f2a9f279e43c1b1089d37c70d9f6032e11948"} Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.388748 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ksr7q\" (UniqueName: \"kubernetes.io/projected/f8123ac7-4c54-4dbf-be97-52221d239894-kube-api-access-ksr7q\") pod \"service-ca-9c57cc56f-4jpx4\" (UID: \"f8123ac7-4c54-4dbf-be97-52221d239894\") " pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.388830 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbcn4\" (UniqueName: \"kubernetes.io/projected/43932d3f-4a3d-4ea0-8d0b-1443c876186b-kube-api-access-bbcn4\") pod \"packageserver-d55dfcdfc-h6kvq\" (UID: \"43932d3f-4a3d-4ea0-8d0b-1443c876186b\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.390033 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftkq9\" (UniqueName: \"kubernetes.io/projected/cbf444c4-dd3b-4cf7-9771-fea9294124c5-kube-api-access-ftkq9\") pod \"route-controller-manager-6576b87f9c-lnmp8\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.399996 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.410424 4815 generic.go:334] "Generic (PLEG): container finished" podID="a34781cf-0b80-4615-913a-8908ca380ef2" containerID="ecd67dbf20ee57215395d185b56206b34973ac2f8aaf056a3db442cfd6201773" exitCode=0 Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.410506 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" event={"ID":"a34781cf-0b80-4615-913a-8908ca380ef2","Type":"ContainerDied","Data":"ecd67dbf20ee57215395d185b56206b34973ac2f8aaf056a3db442cfd6201773"} Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.427224 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c954t\" (UniqueName: \"kubernetes.io/projected/250151de-9f8b-4b4b-af77-b7b35ec0f022-kube-api-access-c954t\") pod \"csi-hostpathplugin-54d4p\" (UID: \"250151de-9f8b-4b4b-af77-b7b35ec0f022\") " pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.437109 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmrrx\" (UniqueName: \"kubernetes.io/projected/3c51b465-c48a-4e6e-b359-7d730bce4980-kube-api-access-mmrrx\") pod \"dns-default-d5gt5\" (UID: \"3c51b465-c48a-4e6e-b359-7d730bce4980\") " pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.442206 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2phf\" (UniqueName: \"kubernetes.io/projected/cfe33ba6-894d-4702-b0b9-43abbbe4d6d7-kube-api-access-d2phf\") pod \"ingress-canary-rhtl7\" (UID: \"cfe33ba6-894d-4702-b0b9-43abbbe4d6d7\") " pod="openshift-ingress-canary/ingress-canary-rhtl7" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.466919 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" event={"ID":"afd36151-d969-4642-b6c4-c6b3b5f65476","Type":"ContainerStarted","Data":"79119808cc81d1589226f0fdce8ea0c82a36b9d45e03687ca80a987018e221dc"} Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.471439 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgwq2\" (UniqueName: \"kubernetes.io/projected/9a9c7304-7002-4b17-9404-d1f17caea8b3-kube-api-access-lgwq2\") pod \"machine-config-controller-84d6567774-zjxb9\" (UID: \"9a9c7304-7002-4b17-9404-d1f17caea8b3\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.478661 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" event={"ID":"1615b53d-0e1b-4d6a-a369-1a614711721c","Type":"ContainerStarted","Data":"e763e78c7e80cf7ddd7fbee9ae46e02d31ab61c5b4edfb40ce94ce6fc42bc0c9"} Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.479479 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.480471 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.482697 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.521783 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b2cadb7f-14cf-4391-a43a-801a1ade7254-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-szts7\" (UID: \"b2cadb7f-14cf-4391-a43a-801a1ade7254\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.523225 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnssk\" (UniqueName: \"kubernetes.io/projected/52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b-kube-api-access-wnssk\") pod \"package-server-manager-789f6589d5-gnsgl\" (UID: \"52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.526459 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:10 crc kubenswrapper[4815]: E1205 09:09:10.529159 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:11.029146515 +0000 UTC m=+149.907753352 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.575747 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.579747 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr7q5\" (UniqueName: \"kubernetes.io/projected/ec745360-4a3f-4af0-ab12-588ef1345804-kube-api-access-tr7q5\") pod \"multus-admission-controller-857f4d67dd-cgjsg\" (UID: \"ec745360-4a3f-4af0-ab12-588ef1345804\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.581110 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.586801 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jj8r\" (UniqueName: \"kubernetes.io/projected/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-kube-api-access-4jj8r\") pod \"marketplace-operator-79b997595-dgk4c\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.587181 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:10 crc kubenswrapper[4815]: E1205 09:09:10.588022 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:11.0879997 +0000 UTC m=+149.966606537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.635221 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq29x\" (UniqueName: \"kubernetes.io/projected/90e530b4-0406-491b-8211-f40acd36696b-kube-api-access-dq29x\") pod \"catalog-operator-68c6474976-6q2w4\" (UID: \"90e530b4-0406-491b-8211-f40acd36696b\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.636038 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhdxx\" (UniqueName: \"kubernetes.io/projected/fb48516f-711a-4131-a555-de5ee6e96ae1-kube-api-access-dhdxx\") pod \"control-plane-machine-set-operator-78cbb6b69f-h89pl\" (UID: \"fb48516f-711a-4131-a555-de5ee6e96ae1\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.636111 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.636304 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkhcx\" (UniqueName: \"kubernetes.io/projected/b63f1c30-9543-4574-bb51-eb7fb474d495-kube-api-access-kkhcx\") pod \"service-ca-operator-777779d784-47vsq\" (UID: \"b63f1c30-9543-4574-bb51-eb7fb474d495\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.636757 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.636850 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.637308 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rhtl7" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.668172 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45j9d\" (UniqueName: \"kubernetes.io/projected/4c34a363-c60d-4d61-a2ea-4d93a54862ba-kube-api-access-45j9d\") pod \"machine-config-server-kj46k\" (UID: \"4c34a363-c60d-4d61-a2ea-4d93a54862ba\") " pod="openshift-machine-config-operator/machine-config-server-kj46k" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.672856 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpg5k\" (UniqueName: \"kubernetes.io/projected/fcbaef71-6da7-4bf7-9ecf-1438e91571b6-kube-api-access-wpg5k\") pod \"migrator-59844c95c7-tmqvz\" (UID: \"fcbaef71-6da7-4bf7-9ecf-1438e91571b6\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.690097 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfccw\" (UniqueName: \"kubernetes.io/projected/025c3afb-3b9a-4fe8-a11c-50b309374464-kube-api-access-gfccw\") pod \"kube-storage-version-migrator-operator-b67b599dd-8vfp6\" (UID: \"025c3afb-3b9a-4fe8-a11c-50b309374464\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.693595 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:10 crc kubenswrapper[4815]: E1205 09:09:10.694073 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:11.194055951 +0000 UTC m=+150.072662788 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.694831 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-54d4p" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.712774 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d30f61cc-7d09-4ae2-86c8-7ec8a168dce8-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-6vqd5\" (UID: \"d30f61cc-7d09-4ae2-86c8-7ec8a168dce8\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.751764 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/36964d20-077c-4fa4-9672-50b92f34a949-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-k6brj\" (UID: \"36964d20-077c-4fa4-9672-50b92f34a949\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.752083 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.755516 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.756015 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.756780 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.766973 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.769625 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2rg4\" (UniqueName: \"kubernetes.io/projected/b78558ed-6530-417a-a88e-a17d601593b0-kube-api-access-r2rg4\") pod \"collect-profiles-29415420-98n6w\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.772152 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5p2pc\" (UniqueName: \"kubernetes.io/projected/59496318-7ad1-4c38-ab56-b51a0ba983dc-kube-api-access-5p2pc\") pod \"olm-operator-6b444d44fb-ctjqr\" (UID: \"59496318-7ad1-4c38-ab56-b51a0ba983dc\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.772455 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.773346 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2427\" (UniqueName: \"kubernetes.io/projected/a4d093b9-c855-48b0-9d44-c43fe89793bb-kube-api-access-x2427\") pod \"oauth-openshift-558db77b4-6sjd8\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.803106 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8hsj\" (UniqueName: \"kubernetes.io/projected/5e9fb212-7595-49b3-8134-4257663737b6-kube-api-access-j8hsj\") pod \"router-default-5444994796-v2lvl\" (UID: \"5e9fb212-7595-49b3-8134-4257663737b6\") " pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.803393 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.804602 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:10 crc kubenswrapper[4815]: E1205 09:09:10.805054 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:11.305034889 +0000 UTC m=+150.183641726 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.817712 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s4ljm\" (UniqueName: \"kubernetes.io/projected/59b3eb4b-02f1-40f2-abb5-b1e0939ce720-kube-api-access-s4ljm\") pod \"machine-config-operator-74547568cd-pzl5s\" (UID: \"59b3eb4b-02f1-40f2-abb5-b1e0939ce720\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.829756 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.854676 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.872286 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.878587 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.885825 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.904803 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.905823 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:10 crc kubenswrapper[4815]: E1205 09:09:10.906294 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:11.406261436 +0000 UTC m=+150.284868273 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.910355 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.910799 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk"] Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.919313 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-kj46k" Dec 05 09:09:10 crc kubenswrapper[4815]: I1205 09:09:10.923871 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-fvz7z"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.008814 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.023481 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.083640 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-ltjzv"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.084150 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.085252 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.085621 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:11.585603942 +0000 UTC m=+150.464210779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.194940 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.196761 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:11.696741776 +0000 UTC m=+150.575348613 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.298066 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.298254 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:11.798226511 +0000 UTC m=+150.676833348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.298340 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.298650 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:11.798638594 +0000 UTC m=+150.677245431 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:11 crc kubenswrapper[4815]: W1205 09:09:11.353809 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod225d33d1_7da4_4937_a843_7594d7549454.slice/crio-8ca9b72edba338bf69636cf08778796eca29792f669142539885a8a5b646a5a3 WatchSource:0}: Error finding container 8ca9b72edba338bf69636cf08778796eca29792f669142539885a8a5b646a5a3: Status 404 returned error can't find the container with id 8ca9b72edba338bf69636cf08778796eca29792f669142539885a8a5b646a5a3 Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.365604 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-4mvlm"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.375853 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h66pq"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.399576 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.399856 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:11.899841911 +0000 UTC m=+150.778448748 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.399927 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.400172 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:11.90016535 +0000 UTC m=+150.778772187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.479801 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"0273f985eb07abba46b8eb9b48721392cd3bb522bfbebb03c45decb38b0900a0"} Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.490841 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" event={"ID":"225d33d1-7da4-4937-a843-7594d7549454","Type":"ContainerStarted","Data":"8ca9b72edba338bf69636cf08778796eca29792f669142539885a8a5b646a5a3"} Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.502229 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.502555 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:12.002540891 +0000 UTC m=+150.881147718 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.525658 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" event={"ID":"1615b53d-0e1b-4d6a-a369-1a614711721c","Type":"ContainerStarted","Data":"b0e71f2dedd1b762415bd5b94524d7ed5f1ae560d5397115e27de69e459d88f9"} Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.550436 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"678f2d56771358a9831423dd5211a63645b1c01d188642783dfa09a2cccf3fd5"} Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.553130 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-d2j7q"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.584610 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.588452 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"dfbbc9574eb98d1a4e1760d0058daf8a44b84625f420b156e3e8958ac9eb6d7d"} Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.593302 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.602260 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.603501 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.603834 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:12.10381907 +0000 UTC m=+150.982425907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.604957 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" event={"ID":"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4","Type":"ContainerStarted","Data":"0d102af981d3b610156f00cadfe05848301613639e6709f1dbe590bd60d68595"} Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.620049 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" event={"ID":"97c971b0-67a2-456b-ae6b-eb4f2a69630d","Type":"ContainerStarted","Data":"62f924890c9c0f222caaf9b32d25b4d42dda25060f423e35e28a3864c50de146"} Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.646742 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.652637 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4jpx4"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.662373 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-tw877" podStartSLOduration=130.662358996 podStartE2EDuration="2m10.662358996s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:11.661281214 +0000 UTC m=+150.539888051" watchObservedRunningTime="2025-12-05 09:09:11.662358996 +0000 UTC m=+150.540965833" Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.665062 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-wjxhv"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.703461 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq"] Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.705032 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.706141 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:12.206124981 +0000 UTC m=+151.084731808 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.807640 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.808805 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:12.308788841 +0000 UTC m=+151.187395678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:11 crc kubenswrapper[4815]: W1205 09:09:11.846375 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf2e6f1de_e80e_45b3_8caa_ca2c16a027bc.slice/crio-0c4ec57624f759f433ebf45e356867605807218c4ac75fba87d6b2eef2bd536b WatchSource:0}: Error finding container 0c4ec57624f759f433ebf45e356867605807218c4ac75fba87d6b2eef2bd536b: Status 404 returned error can't find the container with id 0c4ec57624f759f433ebf45e356867605807218c4ac75fba87d6b2eef2bd536b Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.911392 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.912295 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:12.412271526 +0000 UTC m=+151.290878363 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:11 crc kubenswrapper[4815]: I1205 09:09:11.916683 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:11 crc kubenswrapper[4815]: E1205 09:09:11.919751 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:12.419711697 +0000 UTC m=+151.298318534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.020395 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:12 crc kubenswrapper[4815]: E1205 09:09:12.020752 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:12.520737619 +0000 UTC m=+151.399344456 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.123225 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:12 crc kubenswrapper[4815]: E1205 09:09:12.123751 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:12.623715989 +0000 UTC m=+151.502322826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.163467 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-ssjtf" podStartSLOduration=131.163445023 podStartE2EDuration="2m11.163445023s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:12.096034274 +0000 UTC m=+150.974641121" watchObservedRunningTime="2025-12-05 09:09:12.163445023 +0000 UTC m=+151.042051860" Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.224753 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:12 crc kubenswrapper[4815]: E1205 09:09:12.225111 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:12.725090921 +0000 UTC m=+151.603697758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.275150 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-54d4p"] Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.302525 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-skb9s"] Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.307054 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7"] Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.328657 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:12 crc kubenswrapper[4815]: E1205 09:09:12.329353 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:12.829340108 +0000 UTC m=+151.707946945 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.389475 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8"] Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.431159 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:12 crc kubenswrapper[4815]: E1205 09:09:12.431875 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:12.931851065 +0000 UTC m=+151.810457902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.620905 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:12 crc kubenswrapper[4815]: E1205 09:09:12.621247 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:13.12122893 +0000 UTC m=+151.999835767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.729066 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:12 crc kubenswrapper[4815]: E1205 09:09:12.731466 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:13.230409515 +0000 UTC m=+152.109016352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.738975 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" event={"ID":"f8123ac7-4c54-4dbf-be97-52221d239894","Type":"ContainerStarted","Data":"12e880b383de95785494446e14e9417c2478c098ae87aafa5546020b8c96d43e"} Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.756687 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" event={"ID":"a34781cf-0b80-4615-913a-8908ca380ef2","Type":"ContainerStarted","Data":"9dd447db6d172701059f6b8171bc6da8a56c13ef89737a10d796b534f07d7561"} Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.758621 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" event={"ID":"43932d3f-4a3d-4ea0-8d0b-1443c876186b","Type":"ContainerStarted","Data":"02cfd40b388320af9761f7ead894475e152b85aabac6ad7a74936bcd9cf86a54"} Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.759275 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-v2lvl" event={"ID":"5e9fb212-7595-49b3-8134-4257663737b6","Type":"ContainerStarted","Data":"b2cc60b851ee79b706d8b3ed811fe1ac2f9176f3235dc0c500d337bb49ed6795"} Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.778509 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" event={"ID":"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec","Type":"ContainerStarted","Data":"2100de616fefb7a2d55787eef7ab3133a1a21679781b82f92575fda1cac53b97"} Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.784915 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" event={"ID":"99bb7807-d522-499e-863e-7be3e10ac6ed","Type":"ContainerStarted","Data":"eb5b2ced14822426945e9cb2ac9e8a9f3a6c4c968643ba54f9d077c5587b20bc"} Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.786676 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" event={"ID":"2211eb63-8b44-48c6-9cca-c032127ddb95","Type":"ContainerStarted","Data":"11c85b1726444a97f0aaced95e392afa5fa305f065a8daffc04170b911830eca"} Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.902253 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:12 crc kubenswrapper[4815]: E1205 09:09:12.906078 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:13.406065191 +0000 UTC m=+152.284672028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.936805 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz"] Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.937361 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" podStartSLOduration=131.937351044 podStartE2EDuration="2m11.937351044s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:12.926975074 +0000 UTC m=+151.805581911" watchObservedRunningTime="2025-12-05 09:09:12.937351044 +0000 UTC m=+151.815957881" Dec 05 09:09:12 crc kubenswrapper[4815]: I1205 09:09:12.949342 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kj46k" event={"ID":"4c34a363-c60d-4d61-a2ea-4d93a54862ba","Type":"ContainerStarted","Data":"8646029a17298ed07b0cb9725c3a330865d611214fd00ae8e573efcffe3da34e"} Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.005183 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:13 crc kubenswrapper[4815]: E1205 09:09:13.005401 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:13.505366941 +0000 UTC m=+152.383973778 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.005460 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:13 crc kubenswrapper[4815]: E1205 09:09:13.005842 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:13.505830935 +0000 UTC m=+152.384437772 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.063194 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-d2j7q" event={"ID":"bf95c26c-10fd-41bb-968a-2b5df4066d29","Type":"ContainerStarted","Data":"de0a88d4076e535b7684a5eae61247dcdf9bf3ca04c95d978beca6abd51cbb43"} Dec 05 09:09:13 crc kubenswrapper[4815]: E1205 09:09:13.117264 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:13.617245196 +0000 UTC m=+152.495852033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.136989 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.137303 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.143457 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" event={"ID":"be45782f-d05a-4f70-8d05-3556a37a9698","Type":"ContainerStarted","Data":"7935194c3aceab696b3e836a8dd2861021cd8fd45f8989eeca2f9b8050139940"} Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.144318 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.144602 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:13 crc kubenswrapper[4815]: E1205 09:09:13.144987 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:13.644975373 +0000 UTC m=+152.523582210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.261688 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:13 crc kubenswrapper[4815]: E1205 09:09:13.261966 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:13.76195267 +0000 UTC m=+152.640559507 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.326188 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4mvlm" event={"ID":"a20f4d0e-9b06-4bc0-a532-a104392a18dd","Type":"ContainerStarted","Data":"48703dca2de3741e9391fbe49991ef173c5047a740578bcfb309290819fa0e09"} Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.330676 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" event={"ID":"036169eb-913d-470b-b749-6583deb8f396","Type":"ContainerStarted","Data":"5de27619c47d116ea3754067b8a9a1cfab543f078cd18721d170bde70d160710"} Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.331317 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" event={"ID":"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc","Type":"ContainerStarted","Data":"0c4ec57624f759f433ebf45e356867605807218c4ac75fba87d6b2eef2bd536b"} Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.331875 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" event={"ID":"cd3f420e-13ec-4e1a-8142-31b0bf64493e","Type":"ContainerStarted","Data":"3f3bda7d1c8eaf06ca9aa6927d147cd9af469101f16f687c9487bcc8496654d1"} Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.559771 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:13 crc kubenswrapper[4815]: E1205 09:09:13.560125 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:14.060110338 +0000 UTC m=+152.938717175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.662075 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:13 crc kubenswrapper[4815]: E1205 09:09:13.662416 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:14.162399937 +0000 UTC m=+153.041006774 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.703527 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s"] Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.779431 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:13 crc kubenswrapper[4815]: E1205 09:09:13.779743 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:14.279725965 +0000 UTC m=+153.158332822 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:13 crc kubenswrapper[4815]: I1205 09:09:13.887342 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:13 crc kubenswrapper[4815]: E1205 09:09:13.887596 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:14.3875819 +0000 UTC m=+153.266188737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.142847 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:14 crc kubenswrapper[4815]: E1205 09:09:14.145758 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:14.645741696 +0000 UTC m=+153.524348533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.247969 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:14 crc kubenswrapper[4815]: E1205 09:09:14.248603 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:14.748587942 +0000 UTC m=+153.627194779 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.340544 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" event={"ID":"97c971b0-67a2-456b-ae6b-eb4f2a69630d","Type":"ContainerStarted","Data":"7e5cc8abf5b5379b04c2d8a1787ee913d5d826049720f54cb425b7deda2167b3"} Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.342037 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" event={"ID":"9190adfa-bf4f-421b-a2d9-21bb94b94334","Type":"ContainerStarted","Data":"d9269954254ebaea4f9c4260ec511b3a24a3d129dfa00e89e85e22a5a06c8764"} Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.343112 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" event={"ID":"59b3eb4b-02f1-40f2-abb5-b1e0939ce720","Type":"ContainerStarted","Data":"6b973b237226aad1208e1957d139ca34417bba4a4b2b0fbf27a4f58b51bb520d"} Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.344466 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-54d4p" event={"ID":"250151de-9f8b-4b4b-af77-b7b35ec0f022","Type":"ContainerStarted","Data":"c8aa538778eea4895b5be6691779cb0e9d50b017605764ee892d37f38e1b5557"} Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.346185 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a7ea13d4ef7e17ddd48dc53d1891737afdfdf938c572773daa505bf24093b6fb"} Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.347234 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" event={"ID":"cbf444c4-dd3b-4cf7-9771-fea9294124c5","Type":"ContainerStarted","Data":"eeafab3202013910bc4f98b4e79bb11ef3df20dcbc0e4fe16bd11b3723618bea"} Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.349555 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.349613 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"2760208f0aed3b1b3a00ed398782cd7c755536581fb3c8d160615e965bcfce48"} Dec 05 09:09:14 crc kubenswrapper[4815]: E1205 09:09:14.349917 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:14.849903492 +0000 UTC m=+153.728510319 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.351881 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-4mvlm" event={"ID":"a20f4d0e-9b06-4bc0-a532-a104392a18dd","Type":"ContainerStarted","Data":"6b2702853121625524018f3873a99ee1416b7f72153c70625f0775b5df41e114"} Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.352300 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.354571 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz" event={"ID":"fcbaef71-6da7-4bf7-9ecf-1438e91571b6","Type":"ContainerStarted","Data":"90b663645d0c22d5088e3489bd82121c887bac776530e884c1db0d7c4a255a42"} Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.359611 4815 patch_prober.go:28] interesting pod/console-operator-58897d9998-4mvlm container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.359650 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4mvlm" podUID="a20f4d0e-9b06-4bc0-a532-a104392a18dd" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.359811 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" event={"ID":"e685ebf0-e53e-4a9d-bb44-9e4aa4bd98d4","Type":"ContainerStarted","Data":"3f2a493f4a8c185c4abe26e414ad5e443822d08d706bd40dbfb86375261ba29d"} Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.366118 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" event={"ID":"b2cadb7f-14cf-4391-a43a-801a1ade7254","Type":"ContainerStarted","Data":"56f92e9003a9b6fbe44dce7ba4178e9c6c98040db2b161fc2dbb335ade4a680a"} Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.369901 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" event={"ID":"1615b53d-0e1b-4d6a-a369-1a614711721c","Type":"ContainerStarted","Data":"3f03f8ca372a8cddf563d292415a9e9006576959d040bb5284493b71e3291eff"} Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.372401 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-4mvlm" podStartSLOduration=134.372390282 podStartE2EDuration="2m14.372390282s" podCreationTimestamp="2025-12-05 09:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:14.365672862 +0000 UTC m=+153.244279689" watchObservedRunningTime="2025-12-05 09:09:14.372390282 +0000 UTC m=+153.250997119" Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.393712 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.406638 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.408348 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rhtl7"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.409037 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-fvz7z" podStartSLOduration=133.409026225 podStartE2EDuration="2m13.409026225s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:14.398776649 +0000 UTC m=+153.277383486" watchObservedRunningTime="2025-12-05 09:09:14.409026225 +0000 UTC m=+153.287633062" Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.448275 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-tnfbl" podStartSLOduration=134.448254404 podStartE2EDuration="2m14.448254404s" podCreationTimestamp="2025-12-05 09:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:14.421220148 +0000 UTC m=+153.299826985" watchObservedRunningTime="2025-12-05 09:09:14.448254404 +0000 UTC m=+153.326861241" Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.449483 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-47vsq"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.450426 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:14 crc kubenswrapper[4815]: E1205 09:09:14.450445 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:14.950429439 +0000 UTC m=+153.829036266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.452797 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:14 crc kubenswrapper[4815]: E1205 09:09:14.453970 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:14.953956344 +0000 UTC m=+153.832563181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.463757 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-d5gt5"] Dec 05 09:09:14 crc kubenswrapper[4815]: W1205 09:09:14.511861 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90e530b4_0406_491b_8211_f40acd36696b.slice/crio-395141b7b8c91c2d5b33e9c6779c395ba74a12242d9edd6dcfcbcbe9f2f0989d WatchSource:0}: Error finding container 395141b7b8c91c2d5b33e9c6779c395ba74a12242d9edd6dcfcbcbe9f2f0989d: Status 404 returned error can't find the container with id 395141b7b8c91c2d5b33e9c6779c395ba74a12242d9edd6dcfcbcbe9f2f0989d Dec 05 09:09:14 crc kubenswrapper[4815]: W1205 09:09:14.516475 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfb48516f_711a_4131_a555_de5ee6e96ae1.slice/crio-7e3dc3f26d6b4c94f0d072c0b1ae68242014a403d4465fdac850b94036140c0c WatchSource:0}: Error finding container 7e3dc3f26d6b4c94f0d072c0b1ae68242014a403d4465fdac850b94036140c0c: Status 404 returned error can't find the container with id 7e3dc3f26d6b4c94f0d072c0b1ae68242014a403d4465fdac850b94036140c0c Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.554927 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:14 crc kubenswrapper[4815]: E1205 09:09:14.555900 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:15.055882062 +0000 UTC m=+153.934488899 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:14 crc kubenswrapper[4815]: W1205 09:09:14.568796 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c51b465_c48a_4e6e_b359_7d730bce4980.slice/crio-2f841528ab627fc6459cb457cfe87454496b8997c442cd0eb771a67f5fa1cd4c WatchSource:0}: Error finding container 2f841528ab627fc6459cb457cfe87454496b8997c442cd0eb771a67f5fa1cd4c: Status 404 returned error can't find the container with id 2f841528ab627fc6459cb457cfe87454496b8997c442cd0eb771a67f5fa1cd4c Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.605707 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.616916 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.629751 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgk4c"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.638091 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.657504 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.658080 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:14 crc kubenswrapper[4815]: E1205 09:09:14.658462 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:15.15844907 +0000 UTC m=+154.037055907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.677091 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.680194 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w"] Dec 05 09:09:14 crc kubenswrapper[4815]: W1205 09:09:14.730631 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36964d20_077c_4fa4_9672_50b92f34a949.slice/crio-16abf648afd0eab80e7356612ddea7b28a7aa80d8c5b530d2bec979374a38a1a WatchSource:0}: Error finding container 16abf648afd0eab80e7356612ddea7b28a7aa80d8c5b530d2bec979374a38a1a: Status 404 returned error can't find the container with id 16abf648afd0eab80e7356612ddea7b28a7aa80d8c5b530d2bec979374a38a1a Dec 05 09:09:14 crc kubenswrapper[4815]: W1205 09:09:14.731113 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd30f61cc_7d09_4ae2_86c8_7ec8a168dce8.slice/crio-e775032b0b5dc51af8a982a5d3630441a9f257a42cc7c7dbd0f09842f586c244 WatchSource:0}: Error finding container e775032b0b5dc51af8a982a5d3630441a9f257a42cc7c7dbd0f09842f586c244: Status 404 returned error can't find the container with id e775032b0b5dc51af8a982a5d3630441a9f257a42cc7c7dbd0f09842f586c244 Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.743833 4815 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-fkstw container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 05 09:09:14 crc kubenswrapper[4815]: [+]log ok Dec 05 09:09:14 crc kubenswrapper[4815]: [+]etcd ok Dec 05 09:09:14 crc kubenswrapper[4815]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 05 09:09:14 crc kubenswrapper[4815]: [-]poststarthook/generic-apiserver-start-informers failed: reason withheld Dec 05 09:09:14 crc kubenswrapper[4815]: [+]poststarthook/max-in-flight-filter ok Dec 05 09:09:14 crc kubenswrapper[4815]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 05 09:09:14 crc kubenswrapper[4815]: [+]poststarthook/openshift.io-StartUserInformer ok Dec 05 09:09:14 crc kubenswrapper[4815]: [+]poststarthook/openshift.io-StartOAuthInformer ok Dec 05 09:09:14 crc kubenswrapper[4815]: [+]poststarthook/openshift.io-StartTokenTimeoutUpdater ok Dec 05 09:09:14 crc kubenswrapper[4815]: livez check failed Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.743885 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" podUID="a34781cf-0b80-4615-913a-8908ca380ef2" containerName="oauth-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.759027 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:14 crc kubenswrapper[4815]: E1205 09:09:14.759471 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:15.259453791 +0000 UTC m=+154.138060638 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:14 crc kubenswrapper[4815]: W1205 09:09:14.770579 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb78558ed_6530_417a_a88e_a17d601593b0.slice/crio-97fb7c0bc6bd72a662f2d2b6df40744c8b4be229f86207a612f90db46b61eb26 WatchSource:0}: Error finding container 97fb7c0bc6bd72a662f2d2b6df40744c8b4be229f86207a612f90db46b61eb26: Status 404 returned error can't find the container with id 97fb7c0bc6bd72a662f2d2b6df40744c8b4be229f86207a612f90db46b61eb26 Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.777682 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-cgjsg"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.800074 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.816247 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjd8"] Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.861279 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:14 crc kubenswrapper[4815]: E1205 09:09:14.861623 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:15.361611316 +0000 UTC m=+154.240218153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:14 crc kubenswrapper[4815]: I1205 09:09:14.991971 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:14 crc kubenswrapper[4815]: E1205 09:09:14.992597 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:15.49257496 +0000 UTC m=+154.371181797 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.144462 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:15 crc kubenswrapper[4815]: E1205 09:09:15.144772 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:15.644758656 +0000 UTC m=+154.523365493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.244804 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:15 crc kubenswrapper[4815]: E1205 09:09:15.245154 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:15.745140359 +0000 UTC m=+154.623747196 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.349371 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:15 crc kubenswrapper[4815]: E1205 09:09:15.349702 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:15.849690646 +0000 UTC m=+154.728297473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.451087 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:15 crc kubenswrapper[4815]: E1205 09:09:15.496641 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:15.996599875 +0000 UTC m=+154.875206712 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.552789 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:15 crc kubenswrapper[4815]: E1205 09:09:15.553606 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:16.053585473 +0000 UTC m=+154.932192310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.591128 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" event={"ID":"9a9c7304-7002-4b17-9404-d1f17caea8b3","Type":"ContainerStarted","Data":"14201fc1967cac7fb76ca6f2b247122e63a75fe89f555b4144947fb885a098d3"} Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.608146 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" event={"ID":"36964d20-077c-4fa4-9672-50b92f34a949","Type":"ContainerStarted","Data":"16abf648afd0eab80e7356612ddea7b28a7aa80d8c5b530d2bec979374a38a1a"} Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.610781 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" event={"ID":"97c971b0-67a2-456b-ae6b-eb4f2a69630d","Type":"ContainerStarted","Data":"d4f1f1c21fd83e6914e1cf0e60dca43439b195678c22ba719a0c177aff87c9a9"} Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.623508 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" event={"ID":"43932d3f-4a3d-4ea0-8d0b-1443c876186b","Type":"ContainerStarted","Data":"78eaf0ea082645eeb7aaeab9651268e35faf30f27b335b31a06eea8e629a5396"} Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.624584 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.653898 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:15 crc kubenswrapper[4815]: E1205 09:09:15.654327 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:16.154301085 +0000 UTC m=+155.032907922 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.666955 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-2w9rk" podStartSLOduration=135.666936003 podStartE2EDuration="2m15.666936003s" podCreationTimestamp="2025-12-05 09:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:15.635773834 +0000 UTC m=+154.514380661" watchObservedRunningTime="2025-12-05 09:09:15.666936003 +0000 UTC m=+154.545542830" Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.667460 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" podStartSLOduration=134.667454088 podStartE2EDuration="2m14.667454088s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:15.66484913 +0000 UTC m=+154.543455967" watchObservedRunningTime="2025-12-05 09:09:15.667454088 +0000 UTC m=+154.546060925" Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.683249 4815 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h6kvq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" start-of-body= Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.683304 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" podUID="43932d3f-4a3d-4ea0-8d0b-1443c876186b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.687916 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" event={"ID":"f8123ac7-4c54-4dbf-be97-52221d239894","Type":"ContainerStarted","Data":"02ff91b320a4a8fc43002ec47077939b021dd667b0ad96161fe22921e90540cd"} Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.713906 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" event={"ID":"036169eb-913d-470b-b749-6583deb8f396","Type":"ContainerStarted","Data":"e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e"} Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.714155 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.726014 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-4jpx4" podStartSLOduration=134.725996913 podStartE2EDuration="2m14.725996913s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:15.724418356 +0000 UTC m=+154.603025203" watchObservedRunningTime="2025-12-05 09:09:15.725996913 +0000 UTC m=+154.604603750" Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.756615 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:15 crc kubenswrapper[4815]: E1205 09:09:15.758124 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:16.25811034 +0000 UTC m=+155.136717187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.766287 4815 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-h66pq container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.766354 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" podUID="036169eb-913d-470b-b749-6583deb8f396" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.782999 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz" event={"ID":"fcbaef71-6da7-4bf7-9ecf-1438e91571b6","Type":"ContainerStarted","Data":"83faad54d9ecc0aad62a62d463c997e94544592b75a2db02ee987c120d8986b3"} Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.784210 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" event={"ID":"2211eb63-8b44-48c6-9cca-c032127ddb95","Type":"ContainerStarted","Data":"88657596429a853a4215c25483fb6add9ba928f35c44a13c5d77b25238e539de"} Dec 05 09:09:15 crc kubenswrapper[4815]: I1205 09:09:15.935381 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:15 crc kubenswrapper[4815]: E1205 09:09:15.944995 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:16.444966921 +0000 UTC m=+155.323573758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.153789 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" event={"ID":"59496318-7ad1-4c38-ab56-b51a0ba983dc","Type":"ContainerStarted","Data":"470f3f70b8fef2ccf39cf80e382cc19384ba12a4cebd21c8024d55543d73510b"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.154884 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:16 crc kubenswrapper[4815]: E1205 09:09:16.155790 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:16.655778355 +0000 UTC m=+155.534385182 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.165247 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-kj46k" event={"ID":"4c34a363-c60d-4d61-a2ea-4d93a54862ba","Type":"ContainerStarted","Data":"1046ac6d215fecfbc86bec07094904e570ba65b5e9bfbc780c2a5f28a26af0d7"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.179601 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" event={"ID":"225d33d1-7da4-4937-a843-7594d7549454","Type":"ContainerStarted","Data":"63fb330c8729db92db0e8a56c4b7cf9bcd030805966936e6d11591d097ff4d16"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.180706 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" event={"ID":"52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b","Type":"ContainerStarted","Data":"3131db0bf80bec136983f3ac96c692a1c4416664e0358b9236db9fa302be07f1"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.181549 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d5gt5" event={"ID":"3c51b465-c48a-4e6e-b359-7d730bce4980","Type":"ContainerStarted","Data":"2f841528ab627fc6459cb457cfe87454496b8997c442cd0eb771a67f5fa1cd4c"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.183581 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" event={"ID":"99bb7807-d522-499e-863e-7be3e10ac6ed","Type":"ContainerStarted","Data":"589f919e7445a5e46b1a16acc727dad1532ae41ff634647310e0a06248a562ab"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.186397 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" event={"ID":"cd3f420e-13ec-4e1a-8142-31b0bf64493e","Type":"ContainerStarted","Data":"7e22e1a33fdb7b86dbc0372e13cf8d2549c623ea19beb867c9a0789ab27ffd0f"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.190653 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" event={"ID":"b63f1c30-9543-4574-bb51-eb7fb474d495","Type":"ContainerStarted","Data":"88b9e91e8afd4acfcc9ca6138626b75c007da176d03f3cce4b69f5ee0a5586ae"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.197910 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" event={"ID":"be45782f-d05a-4f70-8d05-3556a37a9698","Type":"ContainerStarted","Data":"87b06247266fc60cb7edb5a7105e43397e9a778a7246c4bb5000f9d3f285412c"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.212250 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" event={"ID":"ec745360-4a3f-4af0-ab12-588ef1345804","Type":"ContainerStarted","Data":"16aa3bd186f609c78a99d2d5709b9555e70a828e69cf4e2099fc9866bbe6a883"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.213427 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" event={"ID":"a4d093b9-c855-48b0-9d44-c43fe89793bb","Type":"ContainerStarted","Data":"981ced39b5b8476ddc5f4b4299f613f80107500064a1de0353b530ebc483e94b"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.214810 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-d2j7q" event={"ID":"bf95c26c-10fd-41bb-968a-2b5df4066d29","Type":"ContainerStarted","Data":"da63efd2ce136aaafcc6a510282ceefd944b7bbc24ee5b12a81ab2e5a4d203fa"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.217567 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" event={"ID":"9190adfa-bf4f-421b-a2d9-21bb94b94334","Type":"ContainerStarted","Data":"4dc69fe898d84a3f0f2a7b9edcac6873d56c856e3420c1aa729713af97847474"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.219555 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" event={"ID":"b78558ed-6530-417a-a88e-a17d601593b0","Type":"ContainerStarted","Data":"97fb7c0bc6bd72a662f2d2b6df40744c8b4be229f86207a612f90db46b61eb26"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.220904 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" event={"ID":"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc","Type":"ContainerStarted","Data":"376ec8718929d9c019933ea4483efa0105baae802421c7ac7bd871b7b6f48c18"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.233696 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" podStartSLOduration=135.233671437 podStartE2EDuration="2m15.233671437s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:16.178687598 +0000 UTC m=+155.057294435" watchObservedRunningTime="2025-12-05 09:09:16.233671437 +0000 UTC m=+155.112278274" Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.371998 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:16 crc kubenswrapper[4815]: E1205 09:09:16.373214 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:16.873194406 +0000 UTC m=+155.751801243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.381868 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-v2lvl" event={"ID":"5e9fb212-7595-49b3-8134-4257663737b6","Type":"ContainerStarted","Data":"c2811c91fd574c3e24d73310abffe710479eb709d00c878e7f3c448331e910dd"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.395476 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" event={"ID":"025c3afb-3b9a-4fe8-a11c-50b309374464","Type":"ContainerStarted","Data":"2c4fd52abaf39ed2353dfdeff3748c7fba990c31dca55b50af014bbbf43147f3"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.396980 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"78e584d346a9db63e5543d0918f2b0df856a37e7b88c5d7fe483168b032190d0"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.451882 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-kj46k" podStartSLOduration=9.451862191 podStartE2EDuration="9.451862191s" podCreationTimestamp="2025-12-05 09:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:16.234860043 +0000 UTC m=+155.113466890" watchObservedRunningTime="2025-12-05 09:09:16.451862191 +0000 UTC m=+155.330469028" Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.455173 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-kvstg" podStartSLOduration=135.455155469 podStartE2EDuration="2m15.455155469s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:16.451306895 +0000 UTC m=+155.329913752" watchObservedRunningTime="2025-12-05 09:09:16.455155469 +0000 UTC m=+155.333762296" Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.481450 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl" event={"ID":"fb48516f-711a-4131-a555-de5ee6e96ae1","Type":"ContainerStarted","Data":"7e3dc3f26d6b4c94f0d072c0b1ae68242014a403d4465fdac850b94036140c0c"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.483902 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:16 crc kubenswrapper[4815]: E1205 09:09:16.486135 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:16.986123112 +0000 UTC m=+155.864729959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.503138 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-q5b6g" podStartSLOduration=135.503121389 podStartE2EDuration="2m15.503121389s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:16.501015027 +0000 UTC m=+155.379621864" watchObservedRunningTime="2025-12-05 09:09:16.503121389 +0000 UTC m=+155.381728216" Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.526738 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" event={"ID":"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25","Type":"ContainerStarted","Data":"51b0fdaca01f9eb47423f3292fbee8b69e62ba341caef35f45bc55c37833aed9"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.589192 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:16 crc kubenswrapper[4815]: E1205 09:09:16.589464 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:17.089437882 +0000 UTC m=+155.968044739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.617264 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-v2lvl" podStartSLOduration=135.617242781 podStartE2EDuration="2m15.617242781s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:16.613936813 +0000 UTC m=+155.492543650" watchObservedRunningTime="2025-12-05 09:09:16.617242781 +0000 UTC m=+155.495849618" Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.691392 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:16 crc kubenswrapper[4815]: E1205 09:09:16.691872 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:17.191855296 +0000 UTC m=+156.070462143 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.722877 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" event={"ID":"d30f61cc-7d09-4ae2-86c8-7ec8a168dce8","Type":"ContainerStarted","Data":"e775032b0b5dc51af8a982a5d3630441a9f257a42cc7c7dbd0f09842f586c244"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.723812 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-d2j7q" podStartSLOduration=136.723802757 podStartE2EDuration="2m16.723802757s" podCreationTimestamp="2025-12-05 09:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:16.721694094 +0000 UTC m=+155.600300931" watchObservedRunningTime="2025-12-05 09:09:16.723802757 +0000 UTC m=+155.602409594" Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.772516 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" event={"ID":"90e530b4-0406-491b-8211-f40acd36696b","Type":"ContainerStarted","Data":"395141b7b8c91c2d5b33e9c6779c395ba74a12242d9edd6dcfcbcbe9f2f0989d"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.796562 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:16 crc kubenswrapper[4815]: E1205 09:09:16.796867 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:17.296848725 +0000 UTC m=+156.175455562 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.797234 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rhtl7" event={"ID":"cfe33ba6-894d-4702-b0b9-43abbbe4d6d7","Type":"ContainerStarted","Data":"b2c68628673b3860a0c0107a08fa27a6c6bcce58889336afb799488b09d9729d"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.797458 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" podStartSLOduration=135.797448983 podStartE2EDuration="2m15.797448983s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:16.795242878 +0000 UTC m=+155.673849715" watchObservedRunningTime="2025-12-05 09:09:16.797448983 +0000 UTC m=+155.676055820" Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.836500 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-hxnqx" podStartSLOduration=135.836469246 podStartE2EDuration="2m15.836469246s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:16.836019743 +0000 UTC m=+155.714626590" watchObservedRunningTime="2025-12-05 09:09:16.836469246 +0000 UTC m=+155.715076083" Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.872965 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" event={"ID":"a0119b9d-7402-4360-98fa-e5c2b1ecf9ec","Type":"ContainerStarted","Data":"aa03853734a17ff209a226c871ed48603d0cb9a4fc5795b6811d92111afabe93"} Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.874072 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.899311 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:16 crc kubenswrapper[4815]: E1205 09:09:16.899672 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:17.39965758 +0000 UTC m=+156.278264417 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.917842 4815 patch_prober.go:28] interesting pod/console-operator-58897d9998-4mvlm container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.917901 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4mvlm" podUID="a20f4d0e-9b06-4bc0-a532-a104392a18dd" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": dial tcp 10.217.0.22:8443: connect: connection refused" Dec 05 09:09:16 crc kubenswrapper[4815]: I1205 09:09:16.964103 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-wjxhv" podStartSLOduration=135.964088591 podStartE2EDuration="2m15.964088591s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:16.962101741 +0000 UTC m=+155.840708578" watchObservedRunningTime="2025-12-05 09:09:16.964088591 +0000 UTC m=+155.842695428" Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:16.999811 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:17 crc kubenswrapper[4815]: E1205 09:09:17.001664 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:17.50163731 +0000 UTC m=+156.380244147 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.027070 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.101416 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:17 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:17 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:17 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.102018 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.102152 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:17 crc kubenswrapper[4815]: E1205 09:09:17.102335 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:17.602322051 +0000 UTC m=+156.480928888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.153370 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-rhtl7" podStartSLOduration=10.153349462 podStartE2EDuration="10.153349462s" podCreationTimestamp="2025-12-05 09:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:17.00598439 +0000 UTC m=+155.884591317" watchObservedRunningTime="2025-12-05 09:09:17.153349462 +0000 UTC m=+156.031956299" Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.223274 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:17 crc kubenswrapper[4815]: E1205 09:09:17.224234 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:17.724212305 +0000 UTC m=+156.602819142 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.225556 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:17 crc kubenswrapper[4815]: E1205 09:09:17.226045 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:17.726029239 +0000 UTC m=+156.604636076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.332730 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:17 crc kubenswrapper[4815]: E1205 09:09:17.333117 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:17.833100111 +0000 UTC m=+156.711706948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.505790 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:17 crc kubenswrapper[4815]: E1205 09:09:17.506361 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:18.006341915 +0000 UTC m=+156.884948752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.609400 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:17 crc kubenswrapper[4815]: E1205 09:09:17.610375 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:18.110357276 +0000 UTC m=+156.988964113 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.716098 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:17 crc kubenswrapper[4815]: E1205 09:09:17.716655 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:18.216639145 +0000 UTC m=+157.095245982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.823690 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:17 crc kubenswrapper[4815]: E1205 09:09:17.824415 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:18.324396367 +0000 UTC m=+157.203003214 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.924097 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-54d4p" event={"ID":"250151de-9f8b-4b4b-af77-b7b35ec0f022","Type":"ContainerStarted","Data":"059dc31f8577d7f50299997153d0163bd20686ee0e88c4d45086d1043d4b4071"} Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.925109 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:17 crc kubenswrapper[4815]: E1205 09:09:17.925388 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:18.425377116 +0000 UTC m=+157.303983953 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.933263 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d5gt5" event={"ID":"3c51b465-c48a-4e6e-b359-7d730bce4980","Type":"ContainerStarted","Data":"51576d92dafda459394a0def1d72027ea414d19d906385eebcc1939c3277affe"} Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.975881 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" event={"ID":"025c3afb-3b9a-4fe8-a11c-50b309374464","Type":"ContainerStarted","Data":"ae34c4e6be685fa1eb2fd495330c5f731bdc840e1b47102f70c3da82758deeb2"} Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.987091 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" event={"ID":"cbf444c4-dd3b-4cf7-9771-fea9294124c5","Type":"ContainerStarted","Data":"682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04"} Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.988042 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.997014 4815 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-lnmp8 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.997070 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" podUID="cbf444c4-dd3b-4cf7-9771-fea9294124c5" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 05 09:09:17 crc kubenswrapper[4815]: I1205 09:09:17.997454 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rhtl7" event={"ID":"cfe33ba6-894d-4702-b0b9-43abbbe4d6d7","Type":"ContainerStarted","Data":"95242cec94508664fdca87aa05f4047acf964dc2f17ce99ff11395373702f48a"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.006953 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" event={"ID":"d30f61cc-7d09-4ae2-86c8-7ec8a168dce8","Type":"ContainerStarted","Data":"9ee1454cb18c49089323b0f0495dd6b8a2381c577eacf492b2da256b8cc7207d"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.027416 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:18 crc kubenswrapper[4815]: E1205 09:09:18.032077 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:18.532052627 +0000 UTC m=+157.410659464 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.037399 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-8vfp6" podStartSLOduration=137.037379176 podStartE2EDuration="2m17.037379176s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.017088671 +0000 UTC m=+156.895695518" watchObservedRunningTime="2025-12-05 09:09:18.037379176 +0000 UTC m=+156.915986003" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.043351 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:18 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:18 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:18 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.043615 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.044588 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" event={"ID":"90e530b4-0406-491b-8211-f40acd36696b","Type":"ContainerStarted","Data":"9feb8456d1402d72925f67083296cc88cd95bdbe5f9168c09b46160ac3035c78"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.046718 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.053108 4815 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6q2w4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.053155 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" podUID="90e530b4-0406-491b-8211-f40acd36696b" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.054592 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" event={"ID":"a4d093b9-c855-48b0-9d44-c43fe89793bb","Type":"ContainerStarted","Data":"2ab5e49b3488a45e7932a9f0d7452a3e9d80bf00677733341a04ef6b3456c083"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.055287 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.056439 4815 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6sjd8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.056477 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" podUID="a4d093b9-c855-48b0-9d44-c43fe89793bb" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.070673 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" event={"ID":"59496318-7ad1-4c38-ab56-b51a0ba983dc","Type":"ContainerStarted","Data":"787f311605809896f1565023ea238137c8cbec8ed974741418a9121d3ada3e15"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.071418 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.071602 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-6vqd5" podStartSLOduration=137.071584875 podStartE2EDuration="2m17.071584875s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.07038628 +0000 UTC m=+156.948993117" watchObservedRunningTime="2025-12-05 09:09:18.071584875 +0000 UTC m=+156.950191712" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.076022 4815 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-ctjqr container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.076069 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" podUID="59496318-7ad1-4c38-ab56-b51a0ba983dc" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.085553 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" event={"ID":"f2e6f1de-e80e-45b3-8caa-ca2c16a027bc","Type":"ContainerStarted","Data":"908bc27bd6fad4836c73e5322c8bcf14be8ea5bb86902cabad0491512430d959"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.095397 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" event={"ID":"36964d20-077c-4fa4-9672-50b92f34a949","Type":"ContainerStarted","Data":"9bd29c2db6f97b1f84b9c4eebbae71fa08bbd641aeb6fdadfc140c14828b6703"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.118989 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" podStartSLOduration=137.118966027 podStartE2EDuration="2m17.118966027s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.118895085 +0000 UTC m=+156.997501932" watchObservedRunningTime="2025-12-05 09:09:18.118966027 +0000 UTC m=+156.997572864" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.130831 4815 generic.go:334] "Generic (PLEG): container finished" podID="9190adfa-bf4f-421b-a2d9-21bb94b94334" containerID="4dc69fe898d84a3f0f2a7b9edcac6873d56c856e3420c1aa729713af97847474" exitCode=0 Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.131346 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" event={"ID":"9190adfa-bf4f-421b-a2d9-21bb94b94334","Type":"ContainerDied","Data":"4dc69fe898d84a3f0f2a7b9edcac6873d56c856e3420c1aa729713af97847474"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.131404 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.137422 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.139257 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.139306 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.139560 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.139595 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:18 crc kubenswrapper[4815]: E1205 09:09:18.140324 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:18.640310944 +0000 UTC m=+157.518917781 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.141119 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl" event={"ID":"fb48516f-711a-4131-a555-de5ee6e96ae1","Type":"ContainerStarted","Data":"b5459996f98e62405183a0bdc7ecaba4f521c4d45aa8c29e1dd238764d866ee8"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.147765 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.160068 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.160447 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.176766 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-j8nhp" podStartSLOduration=137.1767386 podStartE2EDuration="2m17.1767386s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.176393359 +0000 UTC m=+157.055000206" watchObservedRunningTime="2025-12-05 09:09:18.1767386 +0000 UTC m=+157.055345437" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.177657 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" event={"ID":"59b3eb4b-02f1-40f2-abb5-b1e0939ce720","Type":"ContainerStarted","Data":"39cafad66dbc03b4e73c956c58d1bab25d9f0e8b7346982e07856767d69695cf"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.186133 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" event={"ID":"9a9c7304-7002-4b17-9404-d1f17caea8b3","Type":"ContainerStarted","Data":"bf1f49ba3c38bfa1ee2faccf725ef802f1667d7704501cc6d92580fa480830b8"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.194603 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" event={"ID":"b2cadb7f-14cf-4391-a43a-801a1ade7254","Type":"ContainerStarted","Data":"6a1dd79e81b6c697b66d84d9622cb8d2a7fcce10f5553655c051f00d0f2bf873"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.198604 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" event={"ID":"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25","Type":"ContainerStarted","Data":"83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.199050 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.201863 4815 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dgk4c container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.201906 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" podUID="15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.204463 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz" event={"ID":"fcbaef71-6da7-4bf7-9ecf-1438e91571b6","Type":"ContainerStarted","Data":"3ca6597d86ca3a20228c28f94e4b6e050e0105a5d0c8917cab7e72e58b5c5233"} Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.218727 4815 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-h66pq container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.219046 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" podUID="036169eb-913d-470b-b749-6583deb8f396" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.237069 4815 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h6kvq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" start-of-body= Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.237118 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" podUID="43932d3f-4a3d-4ea0-8d0b-1443c876186b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.30:5443/healthz\": dial tcp 10.217.0.30:5443: connect: connection refused" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.239836 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.242665 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fkstw" Dec 05 09:09:18 crc kubenswrapper[4815]: E1205 09:09:18.243871 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:18.74385177 +0000 UTC m=+157.622458607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.263842 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" podStartSLOduration=137.263799165 podStartE2EDuration="2m17.263799165s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.262821647 +0000 UTC m=+157.141428484" watchObservedRunningTime="2025-12-05 09:09:18.263799165 +0000 UTC m=+157.142406002" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.265574 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" podStartSLOduration=137.265560987 podStartE2EDuration="2m17.265560987s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.226222394 +0000 UTC m=+157.104829231" watchObservedRunningTime="2025-12-05 09:09:18.265560987 +0000 UTC m=+157.144167824" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.294177 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" podStartSLOduration=137.29416282 podStartE2EDuration="2m17.29416282s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.291524212 +0000 UTC m=+157.170131049" watchObservedRunningTime="2025-12-05 09:09:18.29416282 +0000 UTC m=+157.172769657" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.316282 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-k6brj" podStartSLOduration=137.316262509 podStartE2EDuration="2m17.316262509s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.315125395 +0000 UTC m=+157.193732232" watchObservedRunningTime="2025-12-05 09:09:18.316262509 +0000 UTC m=+157.194869356" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.344167 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:18 crc kubenswrapper[4815]: E1205 09:09:18.347510 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:18.847478059 +0000 UTC m=+157.726084896 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.352777 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-szts7" podStartSLOduration=137.352759677 podStartE2EDuration="2m17.352759677s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.351105827 +0000 UTC m=+157.229712664" watchObservedRunningTime="2025-12-05 09:09:18.352759677 +0000 UTC m=+157.231366514" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.447171 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:18 crc kubenswrapper[4815]: E1205 09:09:18.447510 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:18.94747442 +0000 UTC m=+157.826081257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.621813 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:18 crc kubenswrapper[4815]: E1205 09:09:18.622140 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:19.122128937 +0000 UTC m=+158.000735774 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.623578 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tmqvz" podStartSLOduration=137.623564739 podStartE2EDuration="2m17.623564739s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.397587663 +0000 UTC m=+157.276194520" watchObservedRunningTime="2025-12-05 09:09:18.623564739 +0000 UTC m=+157.502171576" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.664448 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" podStartSLOduration=137.664418487 podStartE2EDuration="2m17.664418487s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.663940533 +0000 UTC m=+157.542547370" watchObservedRunningTime="2025-12-05 09:09:18.664418487 +0000 UTC m=+157.543025324" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.666893 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-h89pl" podStartSLOduration=137.666884571 podStartE2EDuration="2m17.666884571s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.622408905 +0000 UTC m=+157.501015742" watchObservedRunningTime="2025-12-05 09:09:18.666884571 +0000 UTC m=+157.545491408" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.732621 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:18 crc kubenswrapper[4815]: E1205 09:09:18.733104 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:19.233085684 +0000 UTC m=+158.111692521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.826177 4815 patch_prober.go:28] interesting pod/apiserver-76f77b778f-2nvwc container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 05 09:09:18 crc kubenswrapper[4815]: [+]log ok Dec 05 09:09:18 crc kubenswrapper[4815]: [+]etcd ok Dec 05 09:09:18 crc kubenswrapper[4815]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 05 09:09:18 crc kubenswrapper[4815]: [+]poststarthook/generic-apiserver-start-informers ok Dec 05 09:09:18 crc kubenswrapper[4815]: [+]poststarthook/max-in-flight-filter ok Dec 05 09:09:18 crc kubenswrapper[4815]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 05 09:09:18 crc kubenswrapper[4815]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 05 09:09:18 crc kubenswrapper[4815]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 05 09:09:18 crc kubenswrapper[4815]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 05 09:09:18 crc kubenswrapper[4815]: [+]poststarthook/project.openshift.io-projectcache ok Dec 05 09:09:18 crc kubenswrapper[4815]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 05 09:09:18 crc kubenswrapper[4815]: [+]poststarthook/openshift.io-startinformers ok Dec 05 09:09:18 crc kubenswrapper[4815]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 05 09:09:18 crc kubenswrapper[4815]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 05 09:09:18 crc kubenswrapper[4815]: livez check failed Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.826254 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" podUID="99bb7807-d522-499e-863e-7be3e10ac6ed" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.843702 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:18 crc kubenswrapper[4815]: E1205 09:09:18.844030 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:19.344001781 +0000 UTC m=+158.222608618 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.944961 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:18 crc kubenswrapper[4815]: E1205 09:09:18.945117 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:19.445093734 +0000 UTC m=+158.323700571 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:18 crc kubenswrapper[4815]: I1205 09:09:18.945240 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:18 crc kubenswrapper[4815]: E1205 09:09:18.945637 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:19.44562471 +0000 UTC m=+158.324231537 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.030821 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:19 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:19 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:19 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.031961 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.046132 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:19 crc kubenswrapper[4815]: E1205 09:09:19.046393 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:19.546377813 +0000 UTC m=+158.424984650 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.148952 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:19 crc kubenswrapper[4815]: E1205 09:09:19.149372 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:19.649355413 +0000 UTC m=+158.527962250 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.230611 4815 patch_prober.go:28] interesting pod/console-operator-58897d9998-4mvlm container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.231054 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4mvlm" podUID="a20f4d0e-9b06-4bc0-a532-a104392a18dd" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.250204 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:19 crc kubenswrapper[4815]: E1205 09:09:19.250549 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:19.750534129 +0000 UTC m=+158.629140956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.250659 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" event={"ID":"ec745360-4a3f-4af0-ab12-588ef1345804","Type":"ContainerStarted","Data":"64c95dbffe4cb3bc8425e0e1e9869a8cf45c9fe13a4e4e01061429646d8a85ce"} Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.250692 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" event={"ID":"ec745360-4a3f-4af0-ab12-588ef1345804","Type":"ContainerStarted","Data":"f8b76124bee306a6fab98033327f0c57db09b8ea811d9ef11589fce5e1a6ed5f"} Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.253271 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" event={"ID":"9a9c7304-7002-4b17-9404-d1f17caea8b3","Type":"ContainerStarted","Data":"74589b15b609f81ba64f44f309452bcbc17744581ba3ba0952069e04e3b022a3"} Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.255353 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" event={"ID":"9190adfa-bf4f-421b-a2d9-21bb94b94334","Type":"ContainerStarted","Data":"06612388915a87d07e7f8e93f7ddfae92ad43ff749eba9df41ea831edfbea2cb"} Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.256976 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" event={"ID":"b78558ed-6530-417a-a88e-a17d601593b0","Type":"ContainerStarted","Data":"e48039f48a5987476a6d576e5e98e3bc774c6a2baa8b0ad29636a6cccc757806"} Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.258836 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" event={"ID":"52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b","Type":"ContainerStarted","Data":"76d5975d92bb096c5abc9c8f30fbb8000609c28813e2ae515d1e2c8255418681"} Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.258861 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" event={"ID":"52a7c9ca-ce91-4d6d-b4e1-e5714a2cef4b","Type":"ContainerStarted","Data":"d88401a410480b3ee7be27643958cf828efed948f4f60429462459e934178bf9"} Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.259216 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.261031 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" event={"ID":"59b3eb4b-02f1-40f2-abb5-b1e0939ce720","Type":"ContainerStarted","Data":"4446bf00449e18a51504a8c8dd7d0f4b269377bb0285faa7545c22a3ed4f980e"} Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.267903 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d5gt5" event={"ID":"3c51b465-c48a-4e6e-b359-7d730bce4980","Type":"ContainerStarted","Data":"7f6ceae4eeca7bb5ea09c83b23b9aeaf30d5c10ac21d084c6ce4c3d4d65d3fdb"} Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.268318 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.269796 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" event={"ID":"b63f1c30-9543-4574-bb51-eb7fb474d495","Type":"ContainerStarted","Data":"c1970d8f898d59888c08a4e2a56984f3d186ec2ef0e29ff0dc0741305726b36a"} Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.272553 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" event={"ID":"225d33d1-7da4-4937-a843-7594d7549454","Type":"ContainerStarted","Data":"88a267a86c58b3c5ef61aeec5113108d5db71d4545b4a4acf0858b1a6fc053ea"} Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.274174 4815 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-ctjqr container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.274215 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" podUID="59496318-7ad1-4c38-ab56-b51a0ba983dc" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.275180 4815 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6sjd8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.276187 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" podUID="a4d093b9-c855-48b0-9d44-c43fe89793bb" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.275908 4815 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dgk4c container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.276391 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" podUID="15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.277200 4815 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-6q2w4 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" start-of-body= Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.277239 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" podUID="90e530b4-0406-491b-8211-f40acd36696b" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.28:8443/healthz\": dial tcp 10.217.0.28:8443: connect: connection refused" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.283989 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" podStartSLOduration=139.283973256 podStartE2EDuration="2m19.283973256s" podCreationTimestamp="2025-12-05 09:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:18.820142609 +0000 UTC m=+157.698749456" watchObservedRunningTime="2025-12-05 09:09:19.283973256 +0000 UTC m=+158.162580093" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.285126 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-cgjsg" podStartSLOduration=138.28511951 podStartE2EDuration="2m18.28511951s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:19.281317377 +0000 UTC m=+158.159924214" watchObservedRunningTime="2025-12-05 09:09:19.28511951 +0000 UTC m=+158.163726347" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.355441 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:19 crc kubenswrapper[4815]: E1205 09:09:19.361673 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:19.861659752 +0000 UTC m=+158.740266589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.399754 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.422357 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zjxb9" podStartSLOduration=138.42233908 podStartE2EDuration="2m18.42233908s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:19.389229044 +0000 UTC m=+158.267835871" watchObservedRunningTime="2025-12-05 09:09:19.42233908 +0000 UTC m=+158.300945917" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.424038 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-47vsq" podStartSLOduration=138.424033471 podStartE2EDuration="2m18.424033471s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:19.331204384 +0000 UTC m=+158.209811211" watchObservedRunningTime="2025-12-05 09:09:19.424033471 +0000 UTC m=+158.302640308" Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.464286 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:19 crc kubenswrapper[4815]: E1205 09:09:19.466235 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:19.966197038 +0000 UTC m=+158.844803875 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.566395 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:19 crc kubenswrapper[4815]: E1205 09:09:19.566712 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:20.066701184 +0000 UTC m=+158.945308021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.666917 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:19 crc kubenswrapper[4815]: E1205 09:09:19.667177 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:20.167159179 +0000 UTC m=+159.045766016 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.767668 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:19 crc kubenswrapper[4815]: E1205 09:09:19.767992 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:20.267979254 +0000 UTC m=+159.146586101 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:19 crc kubenswrapper[4815]: I1205 09:09:19.908798 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:19 crc kubenswrapper[4815]: E1205 09:09:19.909029 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:20.409015728 +0000 UTC m=+159.287622555 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.009852 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.010143 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:20.510130563 +0000 UTC m=+159.388737400 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.036854 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:20 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:20 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:20 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.036941 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.048154 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-ltjzv" podStartSLOduration=139.048137386 podStartE2EDuration="2m19.048137386s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:20.046848617 +0000 UTC m=+158.925455454" watchObservedRunningTime="2025-12-05 09:09:20.048137386 +0000 UTC m=+158.926744223" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.110723 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.111051 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:20.61103584 +0000 UTC m=+159.489642677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.135528 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" podStartSLOduration=140.13448942 podStartE2EDuration="2m20.13448942s" podCreationTimestamp="2025-12-05 09:07:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:20.132023807 +0000 UTC m=+159.010630644" watchObservedRunningTime="2025-12-05 09:09:20.13448942 +0000 UTC m=+159.013096247" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.192101 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.192168 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.212166 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.212513 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:20.712501285 +0000 UTC m=+159.591108122 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.214542 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-pzl5s" podStartSLOduration=139.209258028 podStartE2EDuration="2m19.209258028s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:20.208483475 +0000 UTC m=+159.087090312" watchObservedRunningTime="2025-12-05 09:09:20.209258028 +0000 UTC m=+159.087864865" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.214784 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" podStartSLOduration=139.214778163 podStartE2EDuration="2m19.214778163s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:20.181817251 +0000 UTC m=+159.060424078" watchObservedRunningTime="2025-12-05 09:09:20.214778163 +0000 UTC m=+159.093385000" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.243846 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-d5gt5" podStartSLOduration=13.243829569 podStartE2EDuration="13.243829569s" podCreationTimestamp="2025-12-05 09:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:20.242332394 +0000 UTC m=+159.120939231" watchObservedRunningTime="2025-12-05 09:09:20.243829569 +0000 UTC m=+159.122436406" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.267255 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-4mvlm" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.276184 4815 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h6kvq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.277256 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" podUID="43932d3f-4a3d-4ea0-8d0b-1443c876186b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.30:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.279846 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-54d4p" event={"ID":"250151de-9f8b-4b4b-af77-b7b35ec0f022","Type":"ContainerStarted","Data":"ea5ea3334afa6ce9434e540237b38b19b4114b1131cb108d94692777631ef8ac"} Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.282749 4815 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-ctjqr container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.282809 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" podUID="59496318-7ad1-4c38-ab56-b51a0ba983dc" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.283032 4815 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-skb9s container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.283095 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" podUID="9190adfa-bf4f-421b-a2d9-21bb94b94334" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.283256 4815 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6sjd8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.283281 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" podUID="a4d093b9-c855-48b0-9d44-c43fe89793bb" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.313670 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.314113 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:20.814094094 +0000 UTC m=+159.692700931 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.416475 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.425114 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:20.925096892 +0000 UTC m=+159.803703729 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.451234 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-6q2w4" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.458967 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.517752 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.517851 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.017835837 +0000 UTC m=+159.896442674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.518054 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.518406 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.018391894 +0000 UTC m=+159.896998731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.639051 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.639805 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.640046 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.640507 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.140471433 +0000 UTC m=+160.019078290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.640618 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.643050 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.143030369 +0000 UTC m=+160.021637206 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.741370 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.741748 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.241732851 +0000 UTC m=+160.120339688 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.840175 4815 patch_prober.go:28] interesting pod/console-f9d7485db-d2j7q container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.840258 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-d2j7q" podUID="bf95c26c-10fd-41bb-968a-2b5df4066d29" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.843146 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.843522 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.343506176 +0000 UTC m=+160.222113013 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.879731 4815 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dgk4c container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.879778 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" podUID="15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.879791 4815 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-dgk4c container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.879845 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" podUID="15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.915078 4815 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-ctjqr container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.915140 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" podUID="59496318-7ad1-4c38-ab56-b51a0ba983dc" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.915321 4815 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-ctjqr container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" start-of-body= Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.915378 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" podUID="59496318-7ad1-4c38-ab56-b51a0ba983dc" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.39:8443/healthz\": dial tcp 10.217.0.39:8443: connect: connection refused" Dec 05 09:09:20 crc kubenswrapper[4815]: I1205 09:09:20.944070 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:20 crc kubenswrapper[4815]: E1205 09:09:20.944520 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.444484716 +0000 UTC m=+160.323091553 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.034159 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.038886 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:21 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:21 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:21 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.038933 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.044750 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.045143 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.545127776 +0000 UTC m=+160.423734603 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.089621 4815 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6sjd8 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" start-of-body= Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.089689 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" podUID="a4d093b9-c855-48b0-9d44-c43fe89793bb" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.23:6443/healthz\": dial tcp 10.217.0.23:6443: connect: connection refused" Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.149723 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.149880 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.649853688 +0000 UTC m=+160.528460525 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.150028 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.150378 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.650362493 +0000 UTC m=+160.528969320 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.268403 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.268887 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.768866996 +0000 UTC m=+160.647473833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.269121 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.269402 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.769395031 +0000 UTC m=+160.648001868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.312855 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-54d4p" event={"ID":"250151de-9f8b-4b4b-af77-b7b35ec0f022","Type":"ContainerStarted","Data":"b7462698fcc6203445575c15e315e9b3a3383dd3fd1af73e22920a5b55c2e251"} Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.370084 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.370457 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.870442314 +0000 UTC m=+160.749049151 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.485207 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.486768 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:21.98675315 +0000 UTC m=+160.865359987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.629653 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.630041 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.130027021 +0000 UTC m=+161.008633858 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.637808 4815 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h6kvq container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.30:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.637876 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" podUID="43932d3f-4a3d-4ea0-8d0b-1443c876186b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.30:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.638104 4815 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h6kvq container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.30:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.638173 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" podUID="43932d3f-4a3d-4ea0-8d0b-1443c876186b" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.30:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.731227 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.731538 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.231525698 +0000 UTC m=+161.110132535 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.831895 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.832027 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.332001973 +0000 UTC m=+161.210608810 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.832219 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.832551 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.332539079 +0000 UTC m=+161.211145916 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.933129 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.933317 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.433292553 +0000 UTC m=+161.311899380 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:21 crc kubenswrapper[4815]: I1205 09:09:21.933561 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:21 crc kubenswrapper[4815]: E1205 09:09:21.933903 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.43387969 +0000 UTC m=+161.312486527 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.031861 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:22 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:22 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:22 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.032157 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.034478 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.034702 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.534683995 +0000 UTC m=+161.413290832 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.034781 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.035124 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.535108388 +0000 UTC m=+161.413715225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.149620 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.149965 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.649948491 +0000 UTC m=+161.528555328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.264949 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.265373 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.765351361 +0000 UTC m=+161.643958198 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.365879 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.366226 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.866203437 +0000 UTC m=+161.744810274 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.467603 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.467946 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:22.96793307 +0000 UTC m=+161.846539897 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.568424 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.568584 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.06856019 +0000 UTC m=+161.947167027 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.568726 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.569008 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.068997403 +0000 UTC m=+161.947604240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.670103 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.670286 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.170262461 +0000 UTC m=+162.048869298 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.670338 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.670646 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.170634492 +0000 UTC m=+162.049241329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.771427 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.771583 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.271559641 +0000 UTC m=+162.150166478 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.771696 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.771967 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.271946822 +0000 UTC m=+162.150553719 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.872331 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.872717 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.372703446 +0000 UTC m=+162.251310283 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:22 crc kubenswrapper[4815]: I1205 09:09:22.973232 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:22 crc kubenswrapper[4815]: E1205 09:09:22.973550 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.473538122 +0000 UTC m=+162.352144959 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.041859 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:23 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:23 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:23 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.041907 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.074449 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:23 crc kubenswrapper[4815]: E1205 09:09:23.074788 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.574768009 +0000 UTC m=+162.453374846 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.165469 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.175195 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:23 crc kubenswrapper[4815]: E1205 09:09:23.175547 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.675533323 +0000 UTC m=+162.554140160 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.175792 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-2nvwc" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.276879 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:23 crc kubenswrapper[4815]: E1205 09:09:23.276976 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.776961737 +0000 UTC m=+162.655568574 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.277937 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:23 crc kubenswrapper[4815]: E1205 09:09:23.278561 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.778552365 +0000 UTC m=+162.657159202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.311784 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nntc6"] Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.312718 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.317860 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.344732 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nntc6"] Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.379309 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.379514 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-catalog-content\") pod \"community-operators-nntc6\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.379561 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-utilities\") pod \"community-operators-nntc6\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.379594 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l6ct\" (UniqueName: \"kubernetes.io/projected/201dec63-11da-4e0d-93ae-025de8a76beb-kube-api-access-8l6ct\") pod \"community-operators-nntc6\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.379641 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:09:23 crc kubenswrapper[4815]: E1205 09:09:23.379891 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.879877555 +0000 UTC m=+162.758484392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.409296 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/010248a6-d07e-4e6b-97c8-bda1925d6adb-metrics-certs\") pod \"network-metrics-daemon-sslm2\" (UID: \"010248a6-d07e-4e6b-97c8-bda1925d6adb\") " pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.449547 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.481341 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-catalog-content\") pod \"community-operators-nntc6\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.481389 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-utilities\") pod \"community-operators-nntc6\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.481406 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l6ct\" (UniqueName: \"kubernetes.io/projected/201dec63-11da-4e0d-93ae-025de8a76beb-kube-api-access-8l6ct\") pod \"community-operators-nntc6\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.481437 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.482224 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-catalog-content\") pod \"community-operators-nntc6\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.482424 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-utilities\") pod \"community-operators-nntc6\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:09:23 crc kubenswrapper[4815]: E1205 09:09:23.483023 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:23.98301324 +0000 UTC m=+162.861620077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.493828 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rbpxh"] Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.494785 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.523693 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.563197 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l6ct\" (UniqueName: \"kubernetes.io/projected/201dec63-11da-4e0d-93ae-025de8a76beb-kube-api-access-8l6ct\") pod \"community-operators-nntc6\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.582200 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.582355 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-catalog-content\") pod \"certified-operators-rbpxh\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.582396 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7hkkk\" (UniqueName: \"kubernetes.io/projected/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-kube-api-access-7hkkk\") pod \"certified-operators-rbpxh\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.582443 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-utilities\") pod \"certified-operators-rbpxh\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:09:23 crc kubenswrapper[4815]: E1205 09:09:23.582587 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:24.082570587 +0000 UTC m=+162.961177424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.582634 4815 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-skb9s container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.582661 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" podUID="9190adfa-bf4f-421b-a2d9-21bb94b94334" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.588617 4815 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-skb9s container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.588682 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" podUID="9190adfa-bf4f-421b-a2d9-21bb94b94334" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.616923 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rbpxh"] Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.631475 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.673596 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-sslm2" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.683515 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7hkkk\" (UniqueName: \"kubernetes.io/projected/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-kube-api-access-7hkkk\") pod \"certified-operators-rbpxh\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.683557 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.683598 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-utilities\") pod \"certified-operators-rbpxh\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.683633 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-catalog-content\") pod \"certified-operators-rbpxh\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.684131 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-catalog-content\") pod \"certified-operators-rbpxh\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:09:23 crc kubenswrapper[4815]: E1205 09:09:23.684703 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:24.184691842 +0000 UTC m=+163.063298679 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.684930 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-utilities\") pod \"certified-operators-rbpxh\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.785466 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:23 crc kubenswrapper[4815]: E1205 09:09:23.785747 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:24.285732684 +0000 UTC m=+163.164339521 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.887086 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:23 crc kubenswrapper[4815]: E1205 09:09:23.887556 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:24.387544528 +0000 UTC m=+163.266151355 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:23 crc kubenswrapper[4815]: I1205 09:09:23.988777 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:23 crc kubenswrapper[4815]: E1205 09:09:23.989212 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:24.489167619 +0000 UTC m=+163.367774456 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.032536 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:24 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:24 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:24 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.032603 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.090250 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:24 crc kubenswrapper[4815]: E1205 09:09:24.090686 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:24.590668564 +0000 UTC m=+163.469275401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.138722 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rd8pw"] Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.140143 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.193377 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.193704 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wrcp\" (UniqueName: \"kubernetes.io/projected/469b719a-7093-4818-ad40-87fa357a7e82-kube-api-access-5wrcp\") pod \"community-operators-rd8pw\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.193741 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-catalog-content\") pod \"community-operators-rd8pw\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.193789 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-utilities\") pod \"community-operators-rd8pw\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:09:24 crc kubenswrapper[4815]: E1205 09:09:24.193954 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:24.693937843 +0000 UTC m=+163.572544670 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.232896 4815 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.280500 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7hkkk\" (UniqueName: \"kubernetes.io/projected/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-kube-api-access-7hkkk\") pod \"certified-operators-rbpxh\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.294500 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wrcp\" (UniqueName: \"kubernetes.io/projected/469b719a-7093-4818-ad40-87fa357a7e82-kube-api-access-5wrcp\") pod \"community-operators-rd8pw\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.294552 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-catalog-content\") pod \"community-operators-rd8pw\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.294589 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.294609 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-utilities\") pod \"community-operators-rd8pw\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:09:24 crc kubenswrapper[4815]: E1205 09:09:24.295023 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:24.795003196 +0000 UTC m=+163.673610023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.295087 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-catalog-content\") pod \"community-operators-rd8pw\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.295115 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-utilities\") pod \"community-operators-rd8pw\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.334398 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q7gx9"] Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.335505 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.348364 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rd8pw"] Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.371274 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-54d4p" event={"ID":"250151de-9f8b-4b4b-af77-b7b35ec0f022","Type":"ContainerStarted","Data":"a4d2911eeef70fd2989d5f1a1484a9728eaefb9c7a96a3fe11a3b29f7c21e578"} Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.395203 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.395400 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-utilities\") pod \"certified-operators-q7gx9\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.395451 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p5px\" (UniqueName: \"kubernetes.io/projected/baadfde2-0beb-4d47-859e-bc80b697a7ea-kube-api-access-8p5px\") pod \"certified-operators-q7gx9\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.395542 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-catalog-content\") pod \"certified-operators-q7gx9\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:09:24 crc kubenswrapper[4815]: E1205 09:09:24.395664 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:24.895650116 +0000 UTC m=+163.774256953 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.418872 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wrcp\" (UniqueName: \"kubernetes.io/projected/469b719a-7093-4818-ad40-87fa357a7e82-kube-api-access-5wrcp\") pod \"community-operators-rd8pw\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.424793 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.465959 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.497074 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-catalog-content\") pod \"certified-operators-q7gx9\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.497118 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.497153 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-utilities\") pod \"certified-operators-q7gx9\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.497186 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p5px\" (UniqueName: \"kubernetes.io/projected/baadfde2-0beb-4d47-859e-bc80b697a7ea-kube-api-access-8p5px\") pod \"certified-operators-q7gx9\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.498331 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-catalog-content\") pod \"certified-operators-q7gx9\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:09:24 crc kubenswrapper[4815]: E1205 09:09:24.499790 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:24.9997754 +0000 UTC m=+163.878382237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.499781 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-utilities\") pod \"certified-operators-q7gx9\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.568219 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q7gx9"] Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.598069 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:24 crc kubenswrapper[4815]: E1205 09:09:24.598831 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:25.098811192 +0000 UTC m=+163.977418019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.748733 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:24 crc kubenswrapper[4815]: E1205 09:09:24.749120 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:25.249103583 +0000 UTC m=+164.127710420 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.757070 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p5px\" (UniqueName: \"kubernetes.io/projected/baadfde2-0beb-4d47-859e-bc80b697a7ea-kube-api-access-8p5px\") pod \"certified-operators-q7gx9\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.797555 4815 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-05T09:09:24.232918685Z","Handler":null,"Name":""} Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.810075 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:09:24 crc kubenswrapper[4815]: I1205 09:09:24.950887 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:24 crc kubenswrapper[4815]: E1205 09:09:24.951188 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-05 09:09:25.451164276 +0000 UTC m=+164.329771113 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.054147 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:25 crc kubenswrapper[4815]: E1205 09:09:25.054478 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-05 09:09:25.554464506 +0000 UTC m=+164.433071333 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fm9ss" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.083964 4815 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.083992 4815 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.088635 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:25 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:25 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:25 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.088671 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.209243 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.233439 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-54d4p" podStartSLOduration=18.23342255 podStartE2EDuration="18.23342255s" podCreationTimestamp="2025-12-05 09:09:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:25.2324337 +0000 UTC m=+164.111040537" watchObservedRunningTime="2025-12-05 09:09:25.23342255 +0000 UTC m=+164.112029377" Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.250276 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.312292 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.423388 4815 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.423428 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.425410 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.498132 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-sslm2"] Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.647326 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-skb9s" Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.931687 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-k7nc7"] Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.933093 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.936847 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.970221 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-utilities\") pod \"redhat-marketplace-k7nc7\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.970274 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-catalog-content\") pod \"redhat-marketplace-k7nc7\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:09:25 crc kubenswrapper[4815]: I1205 09:09:25.970295 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljzgn\" (UniqueName: \"kubernetes.io/projected/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-kube-api-access-ljzgn\") pod \"redhat-marketplace-k7nc7\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.033713 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:26 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:26 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:26 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.033793 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.081538 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-utilities\") pod \"redhat-marketplace-k7nc7\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.081593 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-catalog-content\") pod \"redhat-marketplace-k7nc7\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.081613 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljzgn\" (UniqueName: \"kubernetes.io/projected/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-kube-api-access-ljzgn\") pod \"redhat-marketplace-k7nc7\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.082182 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-utilities\") pod \"redhat-marketplace-k7nc7\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.082233 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-catalog-content\") pod \"redhat-marketplace-k7nc7\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.102816 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7nc7"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.241346 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nntc6"] Dec 05 09:09:26 crc kubenswrapper[4815]: W1205 09:09:26.268657 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod201dec63_11da_4e0d_93ae_025de8a76beb.slice/crio-379d6bf4161d46e7b1fffb0b4927456a1c33865b5d1fddc3549e7773a7e73694 WatchSource:0}: Error finding container 379d6bf4161d46e7b1fffb0b4927456a1c33865b5d1fddc3549e7773a7e73694: Status 404 returned error can't find the container with id 379d6bf4161d46e7b1fffb0b4927456a1c33865b5d1fddc3549e7773a7e73694 Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.283525 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljzgn\" (UniqueName: \"kubernetes.io/projected/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-kube-api-access-ljzgn\") pod \"redhat-marketplace-k7nc7\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.309572 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.382384 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fm9ss\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.398341 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-whp68"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.399816 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.412395 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-catalog-content\") pod \"redhat-marketplace-whp68\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.412451 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-utilities\") pod \"redhat-marketplace-whp68\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.412516 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkmrq\" (UniqueName: \"kubernetes.io/projected/4d3dc703-3608-44cd-8993-58eaaef64058-kube-api-access-qkmrq\") pod \"redhat-marketplace-whp68\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.430264 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.440253 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.457124 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.469744 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-sslm2" event={"ID":"010248a6-d07e-4e6b-97c8-bda1925d6adb","Type":"ContainerStarted","Data":"971c51ee0f6925a9ff401d9d8c26574226d040f2921edb2c3fd36d72a7f53f2b"} Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.476604 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-whp68"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.495030 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rbpxh"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.534922 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.535249 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.538045 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkmrq\" (UniqueName: \"kubernetes.io/projected/4d3dc703-3608-44cd-8993-58eaaef64058-kube-api-access-qkmrq\") pod \"redhat-marketplace-whp68\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.538095 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-catalog-content\") pod \"redhat-marketplace-whp68\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.538154 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-utilities\") pod \"redhat-marketplace-whp68\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.538678 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-utilities\") pod \"redhat-marketplace-whp68\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.539819 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-catalog-content\") pod \"redhat-marketplace-whp68\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.595272 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.606550 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-kb5lk"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.607655 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.615584 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q7gx9"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.615876 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.619457 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkmrq\" (UniqueName: \"kubernetes.io/projected/4d3dc703-3608-44cd-8993-58eaaef64058-kube-api-access-qkmrq\") pod \"redhat-marketplace-whp68\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.637594 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rd8pw"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.651343 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-utilities\") pod \"redhat-operators-kb5lk\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.651401 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/605fe386-c558-49ca-b986-b2a8d72e98c6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"605fe386-c558-49ca-b986-b2a8d72e98c6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.651418 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/605fe386-c558-49ca-b986-b2a8d72e98c6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"605fe386-c558-49ca-b986-b2a8d72e98c6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.651448 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4nfq\" (UniqueName: \"kubernetes.io/projected/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-kube-api-access-l4nfq\") pod \"redhat-operators-kb5lk\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.651499 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-catalog-content\") pod \"redhat-operators-kb5lk\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.775529 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-catalog-content\") pod \"redhat-operators-kb5lk\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.775614 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-utilities\") pod \"redhat-operators-kb5lk\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.775648 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/605fe386-c558-49ca-b986-b2a8d72e98c6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"605fe386-c558-49ca-b986-b2a8d72e98c6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.775665 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/605fe386-c558-49ca-b986-b2a8d72e98c6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"605fe386-c558-49ca-b986-b2a8d72e98c6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.775696 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l4nfq\" (UniqueName: \"kubernetes.io/projected/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-kube-api-access-l4nfq\") pod \"redhat-operators-kb5lk\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.775901 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/605fe386-c558-49ca-b986-b2a8d72e98c6-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"605fe386-c558-49ca-b986-b2a8d72e98c6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.776112 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-catalog-content\") pod \"redhat-operators-kb5lk\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.776230 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-utilities\") pod \"redhat-operators-kb5lk\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.780588 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kb5lk"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.797214 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nntc6" event={"ID":"201dec63-11da-4e0d-93ae-025de8a76beb","Type":"ContainerStarted","Data":"379d6bf4161d46e7b1fffb0b4927456a1c33865b5d1fddc3549e7773a7e73694"} Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.829119 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.830133 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.875174 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.885406 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/605fe386-c558-49ca-b986-b2a8d72e98c6-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"605fe386-c558-49ca-b986-b2a8d72e98c6\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.889108 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ec415e5-895c-4771-8e48-3240d3c36eeb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3ec415e5-895c-4771-8e48-3240d3c36eeb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.889204 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ec415e5-895c-4771-8e48-3240d3c36eeb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3ec415e5-895c-4771-8e48-3240d3c36eeb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.891125 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.891447 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.944996 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.961783 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.995190 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4nfq\" (UniqueName: \"kubernetes.io/projected/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-kube-api-access-l4nfq\") pod \"redhat-operators-kb5lk\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.997464 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ec415e5-895c-4771-8e48-3240d3c36eeb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3ec415e5-895c-4771-8e48-3240d3c36eeb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.997635 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ec415e5-895c-4771-8e48-3240d3c36eeb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3ec415e5-895c-4771-8e48-3240d3c36eeb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 09:09:26 crc kubenswrapper[4815]: I1205 09:09:26.998261 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ec415e5-895c-4771-8e48-3240d3c36eeb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"3ec415e5-895c-4771-8e48-3240d3c36eeb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.005790 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jctlm"] Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.038211 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.039279 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jctlm"] Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.058714 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:27 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:27 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:27 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.058767 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.119326 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ec415e5-895c-4771-8e48-3240d3c36eeb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"3ec415e5-895c-4771-8e48-3240d3c36eeb\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.254987 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-utilities\") pod \"redhat-operators-jctlm\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.255047 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-catalog-content\") pod \"redhat-operators-jctlm\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.255104 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2cgp\" (UniqueName: \"kubernetes.io/projected/d911f1b0-5e65-45ee-b2b6-d609bc40356c-kube-api-access-q2cgp\") pod \"redhat-operators-jctlm\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.270809 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.353421 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.359765 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-utilities\") pod \"redhat-operators-jctlm\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.359837 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-catalog-content\") pod \"redhat-operators-jctlm\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.359878 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2cgp\" (UniqueName: \"kubernetes.io/projected/d911f1b0-5e65-45ee-b2b6-d609bc40356c-kube-api-access-q2cgp\") pod \"redhat-operators-jctlm\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.360451 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-catalog-content\") pod \"redhat-operators-jctlm\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.361824 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-utilities\") pod \"redhat-operators-jctlm\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.457075 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2cgp\" (UniqueName: \"kubernetes.io/projected/d911f1b0-5e65-45ee-b2b6-d609bc40356c-kube-api-access-q2cgp\") pod \"redhat-operators-jctlm\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.745789 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.881794 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbpxh" event={"ID":"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d","Type":"ContainerStarted","Data":"dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1"} Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.881843 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbpxh" event={"ID":"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d","Type":"ContainerStarted","Data":"8177ec92aa411db5db7dda89a7ed4591ea1276fde7ae597d6ca37902e560ac68"} Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.891378 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-sslm2" event={"ID":"010248a6-d07e-4e6b-97c8-bda1925d6adb","Type":"ContainerStarted","Data":"3bb77733ffa0e4534c7390dbbd68ec7099df6908d199db00dfe1b5bdc29fe1f6"} Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.932162 4815 generic.go:334] "Generic (PLEG): container finished" podID="201dec63-11da-4e0d-93ae-025de8a76beb" containerID="c2e07e58596989263ae33bf893e9506ca5e7fbc0089de6723765ca74e627a5cb" exitCode=0 Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.932275 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nntc6" event={"ID":"201dec63-11da-4e0d-93ae-025de8a76beb","Type":"ContainerDied","Data":"c2e07e58596989263ae33bf893e9506ca5e7fbc0089de6723765ca74e627a5cb"} Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.950485 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:09:27 crc kubenswrapper[4815]: I1205 09:09:27.991736 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd8pw" event={"ID":"469b719a-7093-4818-ad40-87fa357a7e82","Type":"ContainerStarted","Data":"74e70142247e705f9f996fe3a359ba8d3a175774e67ed52d807b4a7cddc373cd"} Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.021403 4815 generic.go:334] "Generic (PLEG): container finished" podID="b78558ed-6530-417a-a88e-a17d601593b0" containerID="e48039f48a5987476a6d576e5e98e3bc774c6a2baa8b0ad29636a6cccc757806" exitCode=0 Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.021648 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" event={"ID":"b78558ed-6530-417a-a88e-a17d601593b0","Type":"ContainerDied","Data":"e48039f48a5987476a6d576e5e98e3bc774c6a2baa8b0ad29636a6cccc757806"} Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.046341 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:28 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:28 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:28 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.046616 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.083934 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7gx9" event={"ID":"baadfde2-0beb-4d47-859e-bc80b697a7ea","Type":"ContainerStarted","Data":"3c0464d3960c8b18500139d23089fbbaf9310af48a66c9e72907bc1aaf8a9e39"} Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.153265 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.153310 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.153340 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.153397 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.346112 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7nc7"] Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.414962 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fm9ss"] Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.596589 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-whp68"] Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.650474 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-d5gt5" Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.828943 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-kb5lk"] Dec 05 09:09:28 crc kubenswrapper[4815]: W1205 09:09:28.834624 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7fa3908_d40f_43d1_9f59_7ad5c8a9877a.slice/crio-918066a5fbb083fa59a021365643bbc67e58f919304853265961f4cc52450d2a WatchSource:0}: Error finding container 918066a5fbb083fa59a021365643bbc67e58f919304853265961f4cc52450d2a: Status 404 returned error can't find the container with id 918066a5fbb083fa59a021365643bbc67e58f919304853265961f4cc52450d2a Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.861872 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 05 09:09:28 crc kubenswrapper[4815]: I1205 09:09:28.876668 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.028754 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:29 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:29 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:29 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.029067 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.085439 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jctlm"] Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.110108 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"605fe386-c558-49ca-b986-b2a8d72e98c6","Type":"ContainerStarted","Data":"cd6fabcebe6b1211252b02b6ef4c821097466d619b4d6057d0bfa3973d6a87e1"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.126063 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb5lk" event={"ID":"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a","Type":"ContainerStarted","Data":"918066a5fbb083fa59a021365643bbc67e58f919304853265961f4cc52450d2a"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.134652 4815 generic.go:334] "Generic (PLEG): container finished" podID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerID="52d80bc9c6229451f2e24f36468de30d9ed19786d83228e3b2cbe9b82c63b7ce" exitCode=0 Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.135505 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7nc7" event={"ID":"8084d369-d4ee-4dad-bdd3-c8d7ea34199c","Type":"ContainerDied","Data":"52d80bc9c6229451f2e24f36468de30d9ed19786d83228e3b2cbe9b82c63b7ce"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.135532 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7nc7" event={"ID":"8084d369-d4ee-4dad-bdd3-c8d7ea34199c","Type":"ContainerStarted","Data":"752903add4dd69f6eaad5077224a283159d390620ec9a5714742e10e027269af"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.139288 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whp68" event={"ID":"4d3dc703-3608-44cd-8993-58eaaef64058","Type":"ContainerStarted","Data":"63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.139325 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whp68" event={"ID":"4d3dc703-3608-44cd-8993-58eaaef64058","Type":"ContainerStarted","Data":"44fc1ff9db400c6f3f5a4712d10dd20317b886179c8959c347878497e5c16d00"} Dec 05 09:09:29 crc kubenswrapper[4815]: W1205 09:09:29.151223 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd911f1b0_5e65_45ee_b2b6_d609bc40356c.slice/crio-6b8060aa79fafea63ad55104d4e478c64d6a322b1ce33d63c9c03c5aa044886c WatchSource:0}: Error finding container 6b8060aa79fafea63ad55104d4e478c64d6a322b1ce33d63c9c03c5aa044886c: Status 404 returned error can't find the container with id 6b8060aa79fafea63ad55104d4e478c64d6a322b1ce33d63c9c03c5aa044886c Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.151641 4815 generic.go:334] "Generic (PLEG): container finished" podID="baadfde2-0beb-4d47-859e-bc80b697a7ea" containerID="acb2fe51d5c916c6a38796623407e537767394a89e4b6186b451bc609db198de" exitCode=0 Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.151713 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7gx9" event={"ID":"baadfde2-0beb-4d47-859e-bc80b697a7ea","Type":"ContainerDied","Data":"acb2fe51d5c916c6a38796623407e537767394a89e4b6186b451bc609db198de"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.173559 4815 generic.go:334] "Generic (PLEG): container finished" podID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" containerID="dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1" exitCode=0 Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.173671 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbpxh" event={"ID":"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d","Type":"ContainerDied","Data":"dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.213204 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3ec415e5-895c-4771-8e48-3240d3c36eeb","Type":"ContainerStarted","Data":"87718f3d9d2f6a6c1a078845ad1174da92f98d69e37fd515edd066798f126d63"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.245481 4815 generic.go:334] "Generic (PLEG): container finished" podID="469b719a-7093-4818-ad40-87fa357a7e82" containerID="e4baae428c8001a5e0d607fb2fc34d0edbfded41df22b20383d4bf4ea0fdb773" exitCode=0 Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.246058 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd8pw" event={"ID":"469b719a-7093-4818-ad40-87fa357a7e82","Type":"ContainerDied","Data":"e4baae428c8001a5e0d607fb2fc34d0edbfded41df22b20383d4bf4ea0fdb773"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.258710 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" event={"ID":"a6ab936d-3cc8-4c09-869a-50783298666d","Type":"ContainerStarted","Data":"cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.258762 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" event={"ID":"a6ab936d-3cc8-4c09-869a-50783298666d","Type":"ContainerStarted","Data":"cb861f3ef2e5bacf5b59fe8fd46010103a4f29926190e45dd5d90eb20fd3c7e0"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.259801 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.271898 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-sslm2" event={"ID":"010248a6-d07e-4e6b-97c8-bda1925d6adb","Type":"ContainerStarted","Data":"c9fa5d08179fe4eed8aeace8203a1e1ee97bfab5d6cd90080ae92e7ab5270039"} Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.290635 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" podStartSLOduration=148.290620595 podStartE2EDuration="2m28.290620595s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:29.289692738 +0000 UTC m=+168.168299595" watchObservedRunningTime="2025-12-05 09:09:29.290620595 +0000 UTC m=+168.169227432" Dec 05 09:09:29 crc kubenswrapper[4815]: I1205 09:09:29.330149 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-sslm2" podStartSLOduration=148.330131673 podStartE2EDuration="2m28.330131673s" podCreationTimestamp="2025-12-05 09:07:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:29.327129884 +0000 UTC m=+168.205736711" watchObservedRunningTime="2025-12-05 09:09:29.330131673 +0000 UTC m=+168.208738510" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.028756 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:30 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:30 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:30 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.029012 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.189500 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.307905 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b78558ed-6530-417a-a88e-a17d601593b0-secret-volume\") pod \"b78558ed-6530-417a-a88e-a17d601593b0\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.308006 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2rg4\" (UniqueName: \"kubernetes.io/projected/b78558ed-6530-417a-a88e-a17d601593b0-kube-api-access-r2rg4\") pod \"b78558ed-6530-417a-a88e-a17d601593b0\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.308055 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b78558ed-6530-417a-a88e-a17d601593b0-config-volume\") pod \"b78558ed-6530-417a-a88e-a17d601593b0\" (UID: \"b78558ed-6530-417a-a88e-a17d601593b0\") " Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.309075 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b78558ed-6530-417a-a88e-a17d601593b0-config-volume" (OuterVolumeSpecName: "config-volume") pod "b78558ed-6530-417a-a88e-a17d601593b0" (UID: "b78558ed-6530-417a-a88e-a17d601593b0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.310464 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"605fe386-c558-49ca-b986-b2a8d72e98c6","Type":"ContainerStarted","Data":"32381fbfd784383d80d9ffdf527c65cb4cad09b9530be1bcab353d2d851d48de"} Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.329303 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b78558ed-6530-417a-a88e-a17d601593b0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b78558ed-6530-417a-a88e-a17d601593b0" (UID: "b78558ed-6530-417a-a88e-a17d601593b0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.336060 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.336042999 podStartE2EDuration="4.336042999s" podCreationTimestamp="2025-12-05 09:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:30.32899769 +0000 UTC m=+169.207604527" watchObservedRunningTime="2025-12-05 09:09:30.336042999 +0000 UTC m=+169.214649836" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.340261 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b78558ed-6530-417a-a88e-a17d601593b0-kube-api-access-r2rg4" (OuterVolumeSpecName: "kube-api-access-r2rg4") pod "b78558ed-6530-417a-a88e-a17d601593b0" (UID: "b78558ed-6530-417a-a88e-a17d601593b0"). InnerVolumeSpecName "kube-api-access-r2rg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.350543 4815 generic.go:334] "Generic (PLEG): container finished" podID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerID="43c495c9693386a162169551d421d4897c94bcd93464858f2281881723a7209b" exitCode=0 Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.350604 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctlm" event={"ID":"d911f1b0-5e65-45ee-b2b6-d609bc40356c","Type":"ContainerDied","Data":"43c495c9693386a162169551d421d4897c94bcd93464858f2281881723a7209b"} Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.350627 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctlm" event={"ID":"d911f1b0-5e65-45ee-b2b6-d609bc40356c","Type":"ContainerStarted","Data":"6b8060aa79fafea63ad55104d4e478c64d6a322b1ce33d63c9c03c5aa044886c"} Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.432704 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b78558ed-6530-417a-a88e-a17d601593b0-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.432744 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2rg4\" (UniqueName: \"kubernetes.io/projected/b78558ed-6530-417a-a88e-a17d601593b0-kube-api-access-r2rg4\") on node \"crc\" DevicePath \"\"" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.432761 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b78558ed-6530-417a-a88e-a17d601593b0-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.489685 4815 generic.go:334] "Generic (PLEG): container finished" podID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerID="bf07c94210d582f53b847a2eeaf053d2f5cda30b8f655128e29b65ce9a9ac371" exitCode=0 Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.489791 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb5lk" event={"ID":"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a","Type":"ContainerDied","Data":"bf07c94210d582f53b847a2eeaf053d2f5cda30b8f655128e29b65ce9a9ac371"} Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.502390 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3ec415e5-895c-4771-8e48-3240d3c36eeb","Type":"ContainerStarted","Data":"d3684a04dc19f76f236eee407be54521af1dca696a3e1a16ef577ea6f631f8ac"} Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.507827 4815 generic.go:334] "Generic (PLEG): container finished" podID="4d3dc703-3608-44cd-8993-58eaaef64058" containerID="63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca" exitCode=0 Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.507898 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whp68" event={"ID":"4d3dc703-3608-44cd-8993-58eaaef64058","Type":"ContainerDied","Data":"63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca"} Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.529274 4815 patch_prober.go:28] interesting pod/console-f9d7485db-d2j7q container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.529335 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-d2j7q" podUID="bf95c26c-10fd-41bb-968a-2b5df4066d29" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.547855 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" event={"ID":"b78558ed-6530-417a-a88e-a17d601593b0","Type":"ContainerDied","Data":"97fb7c0bc6bd72a662f2d2b6df40744c8b4be229f86207a612f90db46b61eb26"} Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.547904 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="97fb7c0bc6bd72a662f2d2b6df40744c8b4be229f86207a612f90db46b61eb26" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.547987 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.552379 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=4.552366118 podStartE2EDuration="4.552366118s" podCreationTimestamp="2025-12-05 09:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:09:30.542297578 +0000 UTC m=+169.420904435" watchObservedRunningTime="2025-12-05 09:09:30.552366118 +0000 UTC m=+169.430972955" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.657418 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h6kvq" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.883093 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:09:30 crc kubenswrapper[4815]: I1205 09:09:30.942994 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-ctjqr" Dec 05 09:09:31 crc kubenswrapper[4815]: I1205 09:09:31.040150 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:31 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:31 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:31 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:31 crc kubenswrapper[4815]: I1205 09:09:31.040219 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:31 crc kubenswrapper[4815]: I1205 09:09:31.108204 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:09:31 crc kubenswrapper[4815]: I1205 09:09:31.629720 4815 generic.go:334] "Generic (PLEG): container finished" podID="3ec415e5-895c-4771-8e48-3240d3c36eeb" containerID="d3684a04dc19f76f236eee407be54521af1dca696a3e1a16ef577ea6f631f8ac" exitCode=0 Dec 05 09:09:31 crc kubenswrapper[4815]: I1205 09:09:31.629796 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3ec415e5-895c-4771-8e48-3240d3c36eeb","Type":"ContainerDied","Data":"d3684a04dc19f76f236eee407be54521af1dca696a3e1a16ef577ea6f631f8ac"} Dec 05 09:09:31 crc kubenswrapper[4815]: I1205 09:09:31.634415 4815 generic.go:334] "Generic (PLEG): container finished" podID="605fe386-c558-49ca-b986-b2a8d72e98c6" containerID="32381fbfd784383d80d9ffdf527c65cb4cad09b9530be1bcab353d2d851d48de" exitCode=0 Dec 05 09:09:31 crc kubenswrapper[4815]: I1205 09:09:31.636144 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"605fe386-c558-49ca-b986-b2a8d72e98c6","Type":"ContainerDied","Data":"32381fbfd784383d80d9ffdf527c65cb4cad09b9530be1bcab353d2d851d48de"} Dec 05 09:09:32 crc kubenswrapper[4815]: I1205 09:09:32.027067 4815 patch_prober.go:28] interesting pod/router-default-5444994796-v2lvl container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 05 09:09:32 crc kubenswrapper[4815]: [-]has-synced failed: reason withheld Dec 05 09:09:32 crc kubenswrapper[4815]: [+]process-running ok Dec 05 09:09:32 crc kubenswrapper[4815]: healthz check failed Dec 05 09:09:32 crc kubenswrapper[4815]: I1205 09:09:32.027341 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-v2lvl" podUID="5e9fb212-7595-49b3-8134-4257663737b6" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:09:33 crc kubenswrapper[4815]: I1205 09:09:33.177629 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:33 crc kubenswrapper[4815]: I1205 09:09:33.192581 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-v2lvl" Dec 05 09:09:33 crc kubenswrapper[4815]: I1205 09:09:33.937198 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.013835 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.017643 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"605fe386-c558-49ca-b986-b2a8d72e98c6","Type":"ContainerDied","Data":"cd6fabcebe6b1211252b02b6ef4c821097466d619b4d6057d0bfa3973d6a87e1"} Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.017701 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd6fabcebe6b1211252b02b6ef4c821097466d619b4d6057d0bfa3973d6a87e1" Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.043189 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/605fe386-c558-49ca-b986-b2a8d72e98c6-kube-api-access\") pod \"605fe386-c558-49ca-b986-b2a8d72e98c6\" (UID: \"605fe386-c558-49ca-b986-b2a8d72e98c6\") " Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.043304 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/605fe386-c558-49ca-b986-b2a8d72e98c6-kubelet-dir\") pod \"605fe386-c558-49ca-b986-b2a8d72e98c6\" (UID: \"605fe386-c558-49ca-b986-b2a8d72e98c6\") " Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.043523 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/605fe386-c558-49ca-b986-b2a8d72e98c6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "605fe386-c558-49ca-b986-b2a8d72e98c6" (UID: "605fe386-c558-49ca-b986-b2a8d72e98c6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.068977 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.129382 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/605fe386-c558-49ca-b986-b2a8d72e98c6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "605fe386-c558-49ca-b986-b2a8d72e98c6" (UID: "605fe386-c558-49ca-b986-b2a8d72e98c6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.144532 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ec415e5-895c-4771-8e48-3240d3c36eeb-kube-api-access\") pod \"3ec415e5-895c-4771-8e48-3240d3c36eeb\" (UID: \"3ec415e5-895c-4771-8e48-3240d3c36eeb\") " Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.144597 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ec415e5-895c-4771-8e48-3240d3c36eeb-kubelet-dir\") pod \"3ec415e5-895c-4771-8e48-3240d3c36eeb\" (UID: \"3ec415e5-895c-4771-8e48-3240d3c36eeb\") " Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.145073 4815 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/605fe386-c558-49ca-b986-b2a8d72e98c6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.145092 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/605fe386-c558-49ca-b986-b2a8d72e98c6-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.146420 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3ec415e5-895c-4771-8e48-3240d3c36eeb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "3ec415e5-895c-4771-8e48-3240d3c36eeb" (UID: "3ec415e5-895c-4771-8e48-3240d3c36eeb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.153894 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ec415e5-895c-4771-8e48-3240d3c36eeb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "3ec415e5-895c-4771-8e48-3240d3c36eeb" (UID: "3ec415e5-895c-4771-8e48-3240d3c36eeb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.246575 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/3ec415e5-895c-4771-8e48-3240d3c36eeb-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 09:09:34 crc kubenswrapper[4815]: I1205 09:09:34.246858 4815 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/3ec415e5-895c-4771-8e48-3240d3c36eeb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 09:09:35 crc kubenswrapper[4815]: I1205 09:09:35.041444 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"3ec415e5-895c-4771-8e48-3240d3c36eeb","Type":"ContainerDied","Data":"87718f3d9d2f6a6c1a078845ad1174da92f98d69e37fd515edd066798f126d63"} Dec 05 09:09:35 crc kubenswrapper[4815]: I1205 09:09:35.041503 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87718f3d9d2f6a6c1a078845ad1174da92f98d69e37fd515edd066798f126d63" Dec 05 09:09:35 crc kubenswrapper[4815]: I1205 09:09:35.041523 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 05 09:09:38 crc kubenswrapper[4815]: I1205 09:09:38.131192 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:38 crc kubenswrapper[4815]: I1205 09:09:38.131505 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:38 crc kubenswrapper[4815]: I1205 09:09:38.131553 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:38 crc kubenswrapper[4815]: I1205 09:09:38.131559 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:38 crc kubenswrapper[4815]: I1205 09:09:38.131622 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-ssjtf" Dec 05 09:09:38 crc kubenswrapper[4815]: I1205 09:09:38.132178 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:38 crc kubenswrapper[4815]: I1205 09:09:38.132214 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:38 crc kubenswrapper[4815]: I1205 09:09:38.132462 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"65fc987e195d1995d818656d0c138c3c22672210a737ab622cdee7143f09f698"} pod="openshift-console/downloads-7954f5f757-ssjtf" containerMessage="Container download-server failed liveness probe, will be restarted" Dec 05 09:09:38 crc kubenswrapper[4815]: I1205 09:09:38.132603 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" containerID="cri-o://65fc987e195d1995d818656d0c138c3c22672210a737ab622cdee7143f09f698" gracePeriod=2 Dec 05 09:09:39 crc kubenswrapper[4815]: I1205 09:09:39.116745 4815 generic.go:334] "Generic (PLEG): container finished" podID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerID="65fc987e195d1995d818656d0c138c3c22672210a737ab622cdee7143f09f698" exitCode=0 Dec 05 09:09:39 crc kubenswrapper[4815]: I1205 09:09:39.116847 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ssjtf" event={"ID":"5655913f-6305-4b7c-bfef-2795e06e0a29","Type":"ContainerDied","Data":"65fc987e195d1995d818656d0c138c3c22672210a737ab622cdee7143f09f698"} Dec 05 09:09:40 crc kubenswrapper[4815]: I1205 09:09:40.128979 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-ssjtf" event={"ID":"5655913f-6305-4b7c-bfef-2795e06e0a29","Type":"ContainerStarted","Data":"3eb72bc934bf5feefd67e92d3f0f52becbab15f2e97007a591b2966a9b61a224"} Dec 05 09:09:40 crc kubenswrapper[4815]: I1205 09:09:40.539684 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:40 crc kubenswrapper[4815]: I1205 09:09:40.543776 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:09:41 crc kubenswrapper[4815]: I1205 09:09:41.157461 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-ssjtf" Dec 05 09:09:41 crc kubenswrapper[4815]: I1205 09:09:41.157503 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:41 crc kubenswrapper[4815]: I1205 09:09:41.157992 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:42 crc kubenswrapper[4815]: I1205 09:09:42.190956 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:42 crc kubenswrapper[4815]: I1205 09:09:42.191268 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:46 crc kubenswrapper[4815]: I1205 09:09:46.463189 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:09:48 crc kubenswrapper[4815]: I1205 09:09:48.131436 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:48 crc kubenswrapper[4815]: I1205 09:09:48.131527 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:48 crc kubenswrapper[4815]: I1205 09:09:48.131550 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:48 crc kubenswrapper[4815]: I1205 09:09:48.131639 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:49 crc kubenswrapper[4815]: I1205 09:09:49.579603 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 05 09:09:50 crc kubenswrapper[4815]: I1205 09:09:50.192400 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:09:50 crc kubenswrapper[4815]: I1205 09:09:50.192464 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:09:50 crc kubenswrapper[4815]: I1205 09:09:50.929305 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gnsgl" Dec 05 09:09:58 crc kubenswrapper[4815]: I1205 09:09:58.131204 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:58 crc kubenswrapper[4815]: I1205 09:09:58.131806 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:09:58 crc kubenswrapper[4815]: I1205 09:09:58.134635 4815 patch_prober.go:28] interesting pod/downloads-7954f5f757-ssjtf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" start-of-body= Dec 05 09:09:58 crc kubenswrapper[4815]: I1205 09:09:58.134706 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-ssjtf" podUID="5655913f-6305-4b7c-bfef-2795e06e0a29" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.8:8080/\": dial tcp 10.217.0.8:8080: connect: connection refused" Dec 05 09:10:02 crc kubenswrapper[4815]: I1205 09:10:02.908344 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 09:10:02 crc kubenswrapper[4815]: E1205 09:10:02.909666 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ec415e5-895c-4771-8e48-3240d3c36eeb" containerName="pruner" Dec 05 09:10:02 crc kubenswrapper[4815]: I1205 09:10:02.909731 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ec415e5-895c-4771-8e48-3240d3c36eeb" containerName="pruner" Dec 05 09:10:02 crc kubenswrapper[4815]: E1205 09:10:02.909808 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="605fe386-c558-49ca-b986-b2a8d72e98c6" containerName="pruner" Dec 05 09:10:02 crc kubenswrapper[4815]: I1205 09:10:02.909826 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="605fe386-c558-49ca-b986-b2a8d72e98c6" containerName="pruner" Dec 05 09:10:02 crc kubenswrapper[4815]: E1205 09:10:02.909894 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b78558ed-6530-417a-a88e-a17d601593b0" containerName="collect-profiles" Dec 05 09:10:02 crc kubenswrapper[4815]: I1205 09:10:02.909907 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b78558ed-6530-417a-a88e-a17d601593b0" containerName="collect-profiles" Dec 05 09:10:02 crc kubenswrapper[4815]: I1205 09:10:02.910476 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b78558ed-6530-417a-a88e-a17d601593b0" containerName="collect-profiles" Dec 05 09:10:02 crc kubenswrapper[4815]: I1205 09:10:02.910549 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="605fe386-c558-49ca-b986-b2a8d72e98c6" containerName="pruner" Dec 05 09:10:02 crc kubenswrapper[4815]: I1205 09:10:02.910609 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ec415e5-895c-4771-8e48-3240d3c36eeb" containerName="pruner" Dec 05 09:10:02 crc kubenswrapper[4815]: I1205 09:10:02.911733 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 09:10:02 crc kubenswrapper[4815]: I1205 09:10:02.915773 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 05 09:10:02 crc kubenswrapper[4815]: I1205 09:10:02.916221 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 05 09:10:02 crc kubenswrapper[4815]: I1205 09:10:02.936830 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 09:10:03 crc kubenswrapper[4815]: I1205 09:10:03.020706 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 09:10:03 crc kubenswrapper[4815]: I1205 09:10:03.020782 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 09:10:03 crc kubenswrapper[4815]: I1205 09:10:03.122229 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 09:10:03 crc kubenswrapper[4815]: I1205 09:10:03.122429 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 09:10:03 crc kubenswrapper[4815]: I1205 09:10:03.122537 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 09:10:03 crc kubenswrapper[4815]: I1205 09:10:03.150526 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 09:10:03 crc kubenswrapper[4815]: I1205 09:10:03.262727 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.296599 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.298803 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.310513 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.390352 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.390732 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-var-lock\") pod \"installer-9-crc\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.390849 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f986909-b717-447f-be83-09b47b6deb8d-kube-api-access\") pod \"installer-9-crc\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.492076 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.492168 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-var-lock\") pod \"installer-9-crc\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.492199 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f986909-b717-447f-be83-09b47b6deb8d-kube-api-access\") pod \"installer-9-crc\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.492573 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.492607 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-var-lock\") pod \"installer-9-crc\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.576436 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f986909-b717-447f-be83-09b47b6deb8d-kube-api-access\") pod \"installer-9-crc\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:10:07 crc kubenswrapper[4815]: I1205 09:10:07.624088 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:10:08 crc kubenswrapper[4815]: I1205 09:10:08.151885 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-ssjtf" Dec 05 09:10:20 crc kubenswrapper[4815]: I1205 09:10:20.192632 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:10:20 crc kubenswrapper[4815]: I1205 09:10:20.193192 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:10:20 crc kubenswrapper[4815]: I1205 09:10:20.193256 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:10:20 crc kubenswrapper[4815]: I1205 09:10:20.194567 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:10:20 crc kubenswrapper[4815]: I1205 09:10:20.194689 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2" gracePeriod=600 Dec 05 09:10:23 crc kubenswrapper[4815]: I1205 09:10:23.892111 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2" exitCode=0 Dec 05 09:10:23 crc kubenswrapper[4815]: I1205 09:10:23.892160 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2"} Dec 05 09:10:26 crc kubenswrapper[4815]: E1205 09:10:26.081456 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 05 09:10:26 crc kubenswrapper[4815]: E1205 09:10:26.082161 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l4nfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-kb5lk_openshift-marketplace(e7fa3908-d40f-43d1-9f59-7ad5c8a9877a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:10:26 crc kubenswrapper[4815]: E1205 09:10:26.083385 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-kb5lk" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" Dec 05 09:10:26 crc kubenswrapper[4815]: E1205 09:10:26.100784 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 05 09:10:26 crc kubenswrapper[4815]: E1205 09:10:26.101128 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q2cgp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-jctlm_openshift-marketplace(d911f1b0-5e65-45ee-b2b6-d609bc40356c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:10:26 crc kubenswrapper[4815]: E1205 09:10:26.102374 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-jctlm" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" Dec 05 09:10:27 crc kubenswrapper[4815]: E1205 09:10:27.654780 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-jctlm" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" Dec 05 09:10:27 crc kubenswrapper[4815]: E1205 09:10:27.654917 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-kb5lk" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" Dec 05 09:10:27 crc kubenswrapper[4815]: E1205 09:10:27.710573 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 05 09:10:27 crc kubenswrapper[4815]: E1205 09:10:27.711034 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ljzgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-k7nc7_openshift-marketplace(8084d369-d4ee-4dad-bdd3-c8d7ea34199c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:10:27 crc kubenswrapper[4815]: E1205 09:10:27.712304 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-k7nc7" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" Dec 05 09:10:29 crc kubenswrapper[4815]: E1205 09:10:29.385717 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-k7nc7" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" Dec 05 09:10:29 crc kubenswrapper[4815]: E1205 09:10:29.458347 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 05 09:10:29 crc kubenswrapper[4815]: E1205 09:10:29.458530 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7hkkk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rbpxh_openshift-marketplace(aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:10:29 crc kubenswrapper[4815]: E1205 09:10:29.459866 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rbpxh" podUID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" Dec 05 09:10:29 crc kubenswrapper[4815]: E1205 09:10:29.464701 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 05 09:10:29 crc kubenswrapper[4815]: E1205 09:10:29.464813 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qkmrq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-whp68_openshift-marketplace(4d3dc703-3608-44cd-8993-58eaaef64058): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:10:29 crc kubenswrapper[4815]: E1205 09:10:29.466032 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-whp68" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" Dec 05 09:10:31 crc kubenswrapper[4815]: E1205 09:10:31.821416 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rbpxh" podUID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" Dec 05 09:10:31 crc kubenswrapper[4815]: E1205 09:10:31.821662 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-whp68" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" Dec 05 09:10:31 crc kubenswrapper[4815]: E1205 09:10:31.935186 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 05 09:10:31 crc kubenswrapper[4815]: E1205 09:10:31.935592 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8p5px,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-q7gx9_openshift-marketplace(baadfde2-0beb-4d47-859e-bc80b697a7ea): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:10:31 crc kubenswrapper[4815]: E1205 09:10:31.936669 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-q7gx9" podUID="baadfde2-0beb-4d47-859e-bc80b697a7ea" Dec 05 09:10:31 crc kubenswrapper[4815]: E1205 09:10:31.955389 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-q7gx9" podUID="baadfde2-0beb-4d47-859e-bc80b697a7ea" Dec 05 09:10:31 crc kubenswrapper[4815]: E1205 09:10:31.960340 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 05 09:10:31 crc kubenswrapper[4815]: E1205 09:10:31.960811 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5wrcp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-rd8pw_openshift-marketplace(469b719a-7093-4818-ad40-87fa357a7e82): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:10:31 crc kubenswrapper[4815]: E1205 09:10:31.962033 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-rd8pw" podUID="469b719a-7093-4818-ad40-87fa357a7e82" Dec 05 09:10:32 crc kubenswrapper[4815]: E1205 09:10:32.020550 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 05 09:10:32 crc kubenswrapper[4815]: E1205 09:10:32.020931 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8l6ct,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-nntc6_openshift-marketplace(201dec63-11da-4e0d-93ae-025de8a76beb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:10:32 crc kubenswrapper[4815]: E1205 09:10:32.022047 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-nntc6" podUID="201dec63-11da-4e0d-93ae-025de8a76beb" Dec 05 09:10:32 crc kubenswrapper[4815]: I1205 09:10:32.374392 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 05 09:10:32 crc kubenswrapper[4815]: I1205 09:10:32.440257 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 05 09:10:32 crc kubenswrapper[4815]: W1205 09:10:32.449709 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2af7ab6f_bc6a_4ed7_974b_c2bfb7f7efa6.slice/crio-4090e5cbc0460fcdfa77ee3b23cd3dbfa4465391c411140f91e157b466c605ca WatchSource:0}: Error finding container 4090e5cbc0460fcdfa77ee3b23cd3dbfa4465391c411140f91e157b466c605ca: Status 404 returned error can't find the container with id 4090e5cbc0460fcdfa77ee3b23cd3dbfa4465391c411140f91e157b466c605ca Dec 05 09:10:32 crc kubenswrapper[4815]: I1205 09:10:32.952888 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"1ea7fac19f76fee3446c61cf65967c171412fa2c53983770bc915a74ac684a09"} Dec 05 09:10:32 crc kubenswrapper[4815]: I1205 09:10:32.956101 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6","Type":"ContainerStarted","Data":"9ee7454cabf084da3be76486bf9856709aa2932d60151f17528e73930ec09999"} Dec 05 09:10:32 crc kubenswrapper[4815]: I1205 09:10:32.956139 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6","Type":"ContainerStarted","Data":"4090e5cbc0460fcdfa77ee3b23cd3dbfa4465391c411140f91e157b466c605ca"} Dec 05 09:10:32 crc kubenswrapper[4815]: I1205 09:10:32.958017 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f986909-b717-447f-be83-09b47b6deb8d","Type":"ContainerStarted","Data":"9430f26de18c05a11a62c4a06d6a592ea91ac54193d51ef92383556784ab57ca"} Dec 05 09:10:32 crc kubenswrapper[4815]: I1205 09:10:32.958037 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f986909-b717-447f-be83-09b47b6deb8d","Type":"ContainerStarted","Data":"1d96bf9e7b50203d2036f9620c3957e5b7c98ee0599eb6681ea4106dff24226d"} Dec 05 09:10:32 crc kubenswrapper[4815]: E1205 09:10:32.958522 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-nntc6" podUID="201dec63-11da-4e0d-93ae-025de8a76beb" Dec 05 09:10:32 crc kubenswrapper[4815]: E1205 09:10:32.959234 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-rd8pw" podUID="469b719a-7093-4818-ad40-87fa357a7e82" Dec 05 09:10:33 crc kubenswrapper[4815]: I1205 09:10:33.040027 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=26.04000817 podStartE2EDuration="26.04000817s" podCreationTimestamp="2025-12-05 09:10:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:10:33.038140048 +0000 UTC m=+231.916746895" watchObservedRunningTime="2025-12-05 09:10:33.04000817 +0000 UTC m=+231.918615007" Dec 05 09:10:33 crc kubenswrapper[4815]: I1205 09:10:33.058723 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=31.058700702 podStartE2EDuration="31.058700702s" podCreationTimestamp="2025-12-05 09:10:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:10:33.057683243 +0000 UTC m=+231.936290080" watchObservedRunningTime="2025-12-05 09:10:33.058700702 +0000 UTC m=+231.937307539" Dec 05 09:10:33 crc kubenswrapper[4815]: I1205 09:10:33.982567 4815 generic.go:334] "Generic (PLEG): container finished" podID="2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6" containerID="9ee7454cabf084da3be76486bf9856709aa2932d60151f17528e73930ec09999" exitCode=0 Dec 05 09:10:33 crc kubenswrapper[4815]: I1205 09:10:33.982652 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6","Type":"ContainerDied","Data":"9ee7454cabf084da3be76486bf9856709aa2932d60151f17528e73930ec09999"} Dec 05 09:10:35 crc kubenswrapper[4815]: I1205 09:10:35.213915 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 09:10:35 crc kubenswrapper[4815]: I1205 09:10:35.354953 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kubelet-dir\") pod \"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6\" (UID: \"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6\") " Dec 05 09:10:35 crc kubenswrapper[4815]: I1205 09:10:35.355064 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6" (UID: "2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:10:35 crc kubenswrapper[4815]: I1205 09:10:35.355385 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kube-api-access\") pod \"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6\" (UID: \"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6\") " Dec 05 09:10:35 crc kubenswrapper[4815]: I1205 09:10:35.357249 4815 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:35 crc kubenswrapper[4815]: I1205 09:10:35.361933 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6" (UID: "2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:10:35 crc kubenswrapper[4815]: I1205 09:10:35.458359 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:35 crc kubenswrapper[4815]: I1205 09:10:35.993968 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6","Type":"ContainerDied","Data":"4090e5cbc0460fcdfa77ee3b23cd3dbfa4465391c411140f91e157b466c605ca"} Dec 05 09:10:35 crc kubenswrapper[4815]: I1205 09:10:35.994277 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4090e5cbc0460fcdfa77ee3b23cd3dbfa4465391c411140f91e157b466c605ca" Dec 05 09:10:35 crc kubenswrapper[4815]: I1205 09:10:35.994054 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 05 09:10:40 crc kubenswrapper[4815]: I1205 09:10:40.034570 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctlm" event={"ID":"d911f1b0-5e65-45ee-b2b6-d609bc40356c","Type":"ContainerStarted","Data":"8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7"} Dec 05 09:10:40 crc kubenswrapper[4815]: I1205 09:10:40.145022 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjd8"] Dec 05 09:10:41 crc kubenswrapper[4815]: I1205 09:10:41.042826 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb5lk" event={"ID":"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a","Type":"ContainerStarted","Data":"7b3969eeb2c2ac590d775c30bacd72f6a8027ecae7e46b4a437aeb1743b7adda"} Dec 05 09:10:41 crc kubenswrapper[4815]: I1205 09:10:41.045687 4815 generic.go:334] "Generic (PLEG): container finished" podID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerID="8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7" exitCode=0 Dec 05 09:10:41 crc kubenswrapper[4815]: I1205 09:10:41.045717 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctlm" event={"ID":"d911f1b0-5e65-45ee-b2b6-d609bc40356c","Type":"ContainerDied","Data":"8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7"} Dec 05 09:10:42 crc kubenswrapper[4815]: I1205 09:10:42.055365 4815 generic.go:334] "Generic (PLEG): container finished" podID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerID="7b3969eeb2c2ac590d775c30bacd72f6a8027ecae7e46b4a437aeb1743b7adda" exitCode=0 Dec 05 09:10:42 crc kubenswrapper[4815]: I1205 09:10:42.055569 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb5lk" event={"ID":"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a","Type":"ContainerDied","Data":"7b3969eeb2c2ac590d775c30bacd72f6a8027ecae7e46b4a437aeb1743b7adda"} Dec 05 09:10:43 crc kubenswrapper[4815]: I1205 09:10:43.097268 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctlm" event={"ID":"d911f1b0-5e65-45ee-b2b6-d609bc40356c","Type":"ContainerStarted","Data":"431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df"} Dec 05 09:10:43 crc kubenswrapper[4815]: I1205 09:10:43.105766 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb5lk" event={"ID":"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a","Type":"ContainerStarted","Data":"236ca6b864b6e63b6a57d4677aee4d7e6c8fe6116c9aede3d6af76b45a3ec9d8"} Dec 05 09:10:43 crc kubenswrapper[4815]: I1205 09:10:43.123397 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jctlm" podStartSLOduration=4.802295985 podStartE2EDuration="1m17.123378468s" podCreationTimestamp="2025-12-05 09:09:26 +0000 UTC" firstStartedPulling="2025-12-05 09:09:30.382189066 +0000 UTC m=+169.260795913" lastFinishedPulling="2025-12-05 09:10:42.703271569 +0000 UTC m=+241.581878396" observedRunningTime="2025-12-05 09:10:43.122090131 +0000 UTC m=+242.000696998" watchObservedRunningTime="2025-12-05 09:10:43.123378468 +0000 UTC m=+242.001985315" Dec 05 09:10:43 crc kubenswrapper[4815]: I1205 09:10:43.153212 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-kb5lk" podStartSLOduration=4.839923351 podStartE2EDuration="1m17.153194964s" podCreationTimestamp="2025-12-05 09:09:26 +0000 UTC" firstStartedPulling="2025-12-05 09:09:30.501723638 +0000 UTC m=+169.380330475" lastFinishedPulling="2025-12-05 09:10:42.814995251 +0000 UTC m=+241.693602088" observedRunningTime="2025-12-05 09:10:43.146916756 +0000 UTC m=+242.025523593" watchObservedRunningTime="2025-12-05 09:10:43.153194964 +0000 UTC m=+242.031801801" Dec 05 09:10:44 crc kubenswrapper[4815]: I1205 09:10:44.120505 4815 generic.go:334] "Generic (PLEG): container finished" podID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerID="a04066155246b1f7e94c0699b459498b72601def19cc4a9ffeabaf9ea91b5017" exitCode=0 Dec 05 09:10:44 crc kubenswrapper[4815]: I1205 09:10:44.120878 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7nc7" event={"ID":"8084d369-d4ee-4dad-bdd3-c8d7ea34199c","Type":"ContainerDied","Data":"a04066155246b1f7e94c0699b459498b72601def19cc4a9ffeabaf9ea91b5017"} Dec 05 09:10:44 crc kubenswrapper[4815]: I1205 09:10:44.125978 4815 generic.go:334] "Generic (PLEG): container finished" podID="baadfde2-0beb-4d47-859e-bc80b697a7ea" containerID="870d461dc7121d9c9d713dd5d5432e0d7aa0a0bfb6307ef895fbcc17b28d4c89" exitCode=0 Dec 05 09:10:44 crc kubenswrapper[4815]: I1205 09:10:44.126061 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7gx9" event={"ID":"baadfde2-0beb-4d47-859e-bc80b697a7ea","Type":"ContainerDied","Data":"870d461dc7121d9c9d713dd5d5432e0d7aa0a0bfb6307ef895fbcc17b28d4c89"} Dec 05 09:10:46 crc kubenswrapper[4815]: I1205 09:10:46.137451 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7nc7" event={"ID":"8084d369-d4ee-4dad-bdd3-c8d7ea34199c","Type":"ContainerStarted","Data":"e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f"} Dec 05 09:10:46 crc kubenswrapper[4815]: I1205 09:10:46.139192 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7gx9" event={"ID":"baadfde2-0beb-4d47-859e-bc80b697a7ea","Type":"ContainerStarted","Data":"cb3262b05280e6bbe85ec684136477a37cbb9f86b653f592e20d20a9a727b8e6"} Dec 05 09:10:46 crc kubenswrapper[4815]: I1205 09:10:46.162805 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-k7nc7" podStartSLOduration=4.868761565 podStartE2EDuration="1m21.162791012s" podCreationTimestamp="2025-12-05 09:09:25 +0000 UTC" firstStartedPulling="2025-12-05 09:09:29.136566423 +0000 UTC m=+168.015173260" lastFinishedPulling="2025-12-05 09:10:45.43059587 +0000 UTC m=+244.309202707" observedRunningTime="2025-12-05 09:10:46.161948287 +0000 UTC m=+245.040555124" watchObservedRunningTime="2025-12-05 09:10:46.162791012 +0000 UTC m=+245.041397849" Dec 05 09:10:46 crc kubenswrapper[4815]: I1205 09:10:46.186631 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q7gx9" podStartSLOduration=4.812980769 podStartE2EDuration="1m22.186615538s" podCreationTimestamp="2025-12-05 09:09:24 +0000 UTC" firstStartedPulling="2025-12-05 09:09:28.086626144 +0000 UTC m=+166.965232981" lastFinishedPulling="2025-12-05 09:10:45.460260913 +0000 UTC m=+244.338867750" observedRunningTime="2025-12-05 09:10:46.185322791 +0000 UTC m=+245.063929628" watchObservedRunningTime="2025-12-05 09:10:46.186615538 +0000 UTC m=+245.065222375" Dec 05 09:10:46 crc kubenswrapper[4815]: I1205 09:10:46.311778 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:10:46 crc kubenswrapper[4815]: I1205 09:10:46.311838 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:10:47 crc kubenswrapper[4815]: I1205 09:10:47.155163 4815 generic.go:334] "Generic (PLEG): container finished" podID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" containerID="5ea687eb75a04e26956a84004a193937e60aa9219245d462e6aa8d026d548cdb" exitCode=0 Dec 05 09:10:47 crc kubenswrapper[4815]: I1205 09:10:47.155781 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbpxh" event={"ID":"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d","Type":"ContainerDied","Data":"5ea687eb75a04e26956a84004a193937e60aa9219245d462e6aa8d026d548cdb"} Dec 05 09:10:47 crc kubenswrapper[4815]: I1205 09:10:47.160910 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nntc6" event={"ID":"201dec63-11da-4e0d-93ae-025de8a76beb","Type":"ContainerStarted","Data":"f5d15f79a48da0117d5b75e7d92e9451a6d191594f117b1d95195379fc6587f2"} Dec 05 09:10:47 crc kubenswrapper[4815]: I1205 09:10:47.272098 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:10:47 crc kubenswrapper[4815]: I1205 09:10:47.272203 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:10:47 crc kubenswrapper[4815]: I1205 09:10:47.747219 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:10:47 crc kubenswrapper[4815]: I1205 09:10:47.747264 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:10:47 crc kubenswrapper[4815]: I1205 09:10:47.789827 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-k7nc7" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerName="registry-server" probeResult="failure" output=< Dec 05 09:10:47 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 09:10:47 crc kubenswrapper[4815]: > Dec 05 09:10:48 crc kubenswrapper[4815]: I1205 09:10:48.165980 4815 generic.go:334] "Generic (PLEG): container finished" podID="201dec63-11da-4e0d-93ae-025de8a76beb" containerID="f5d15f79a48da0117d5b75e7d92e9451a6d191594f117b1d95195379fc6587f2" exitCode=0 Dec 05 09:10:48 crc kubenswrapper[4815]: I1205 09:10:48.166029 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nntc6" event={"ID":"201dec63-11da-4e0d-93ae-025de8a76beb","Type":"ContainerDied","Data":"f5d15f79a48da0117d5b75e7d92e9451a6d191594f117b1d95195379fc6587f2"} Dec 05 09:10:48 crc kubenswrapper[4815]: I1205 09:10:48.314417 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-kb5lk" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerName="registry-server" probeResult="failure" output=< Dec 05 09:10:48 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 09:10:48 crc kubenswrapper[4815]: > Dec 05 09:10:48 crc kubenswrapper[4815]: I1205 09:10:48.791874 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-jctlm" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerName="registry-server" probeResult="failure" output=< Dec 05 09:10:48 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 09:10:48 crc kubenswrapper[4815]: > Dec 05 09:10:53 crc kubenswrapper[4815]: I1205 09:10:53.198980 4815 generic.go:334] "Generic (PLEG): container finished" podID="4d3dc703-3608-44cd-8993-58eaaef64058" containerID="e334c47c05dd792a9c646cbfa68e50b1e066559cf058b770d571df1d0e4a3931" exitCode=0 Dec 05 09:10:53 crc kubenswrapper[4815]: I1205 09:10:53.199099 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whp68" event={"ID":"4d3dc703-3608-44cd-8993-58eaaef64058","Type":"ContainerDied","Data":"e334c47c05dd792a9c646cbfa68e50b1e066559cf058b770d571df1d0e4a3931"} Dec 05 09:10:54 crc kubenswrapper[4815]: I1205 09:10:54.810714 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:10:54 crc kubenswrapper[4815]: I1205 09:10:54.811629 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:10:55 crc kubenswrapper[4815]: I1205 09:10:55.018087 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:10:55 crc kubenswrapper[4815]: I1205 09:10:55.252021 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:10:55 crc kubenswrapper[4815]: I1205 09:10:55.658784 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q7gx9"] Dec 05 09:10:56 crc kubenswrapper[4815]: I1205 09:10:56.358333 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:10:56 crc kubenswrapper[4815]: I1205 09:10:56.399828 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:10:57 crc kubenswrapper[4815]: I1205 09:10:57.232447 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nntc6" event={"ID":"201dec63-11da-4e0d-93ae-025de8a76beb","Type":"ContainerStarted","Data":"a5974d90794dd4068b02899a36ee5661556df8ec420b6f0ea92704a153f3018d"} Dec 05 09:10:57 crc kubenswrapper[4815]: I1205 09:10:57.232838 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-q7gx9" podUID="baadfde2-0beb-4d47-859e-bc80b697a7ea" containerName="registry-server" containerID="cri-o://cb3262b05280e6bbe85ec684136477a37cbb9f86b653f592e20d20a9a727b8e6" gracePeriod=2 Dec 05 09:10:57 crc kubenswrapper[4815]: I1205 09:10:57.253019 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nntc6" podStartSLOduration=7.07740335 podStartE2EDuration="1m34.252997496s" podCreationTimestamp="2025-12-05 09:09:23 +0000 UTC" firstStartedPulling="2025-12-05 09:09:27.950231669 +0000 UTC m=+166.828838506" lastFinishedPulling="2025-12-05 09:10:55.125825815 +0000 UTC m=+254.004432652" observedRunningTime="2025-12-05 09:10:57.252482691 +0000 UTC m=+256.131089548" watchObservedRunningTime="2025-12-05 09:10:57.252997496 +0000 UTC m=+256.131604333" Dec 05 09:10:57 crc kubenswrapper[4815]: I1205 09:10:57.315473 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:10:57 crc kubenswrapper[4815]: I1205 09:10:57.366620 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:10:57 crc kubenswrapper[4815]: I1205 09:10:57.794457 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:10:57 crc kubenswrapper[4815]: I1205 09:10:57.835474 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:10:58 crc kubenswrapper[4815]: I1205 09:10:58.238910 4815 generic.go:334] "Generic (PLEG): container finished" podID="baadfde2-0beb-4d47-859e-bc80b697a7ea" containerID="cb3262b05280e6bbe85ec684136477a37cbb9f86b653f592e20d20a9a727b8e6" exitCode=0 Dec 05 09:10:58 crc kubenswrapper[4815]: I1205 09:10:58.239003 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7gx9" event={"ID":"baadfde2-0beb-4d47-859e-bc80b697a7ea","Type":"ContainerDied","Data":"cb3262b05280e6bbe85ec684136477a37cbb9f86b653f592e20d20a9a727b8e6"} Dec 05 09:10:58 crc kubenswrapper[4815]: I1205 09:10:58.859381 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.026927 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8p5px\" (UniqueName: \"kubernetes.io/projected/baadfde2-0beb-4d47-859e-bc80b697a7ea-kube-api-access-8p5px\") pod \"baadfde2-0beb-4d47-859e-bc80b697a7ea\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.027025 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-catalog-content\") pod \"baadfde2-0beb-4d47-859e-bc80b697a7ea\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.027083 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-utilities\") pod \"baadfde2-0beb-4d47-859e-bc80b697a7ea\" (UID: \"baadfde2-0beb-4d47-859e-bc80b697a7ea\") " Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.028184 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-utilities" (OuterVolumeSpecName: "utilities") pod "baadfde2-0beb-4d47-859e-bc80b697a7ea" (UID: "baadfde2-0beb-4d47-859e-bc80b697a7ea"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.033419 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baadfde2-0beb-4d47-859e-bc80b697a7ea-kube-api-access-8p5px" (OuterVolumeSpecName: "kube-api-access-8p5px") pod "baadfde2-0beb-4d47-859e-bc80b697a7ea" (UID: "baadfde2-0beb-4d47-859e-bc80b697a7ea"). InnerVolumeSpecName "kube-api-access-8p5px". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.089959 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "baadfde2-0beb-4d47-859e-bc80b697a7ea" (UID: "baadfde2-0beb-4d47-859e-bc80b697a7ea"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.131168 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.131436 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/baadfde2-0beb-4d47-859e-bc80b697a7ea-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.131559 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8p5px\" (UniqueName: \"kubernetes.io/projected/baadfde2-0beb-4d47-859e-bc80b697a7ea-kube-api-access-8p5px\") on node \"crc\" DevicePath \"\"" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.248533 4815 generic.go:334] "Generic (PLEG): container finished" podID="469b719a-7093-4818-ad40-87fa357a7e82" containerID="6edb26cdf605adc1b660738bb7f2ce1c8b98126cf1e3b6daeb99f1f0ca495a1c" exitCode=0 Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.248619 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd8pw" event={"ID":"469b719a-7093-4818-ad40-87fa357a7e82","Type":"ContainerDied","Data":"6edb26cdf605adc1b660738bb7f2ce1c8b98126cf1e3b6daeb99f1f0ca495a1c"} Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.255444 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whp68" event={"ID":"4d3dc703-3608-44cd-8993-58eaaef64058","Type":"ContainerStarted","Data":"89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26"} Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.258991 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q7gx9" event={"ID":"baadfde2-0beb-4d47-859e-bc80b697a7ea","Type":"ContainerDied","Data":"3c0464d3960c8b18500139d23089fbbaf9310af48a66c9e72907bc1aaf8a9e39"} Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.259049 4815 scope.go:117] "RemoveContainer" containerID="cb3262b05280e6bbe85ec684136477a37cbb9f86b653f592e20d20a9a727b8e6" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.259212 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q7gx9" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.265827 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbpxh" event={"ID":"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d","Type":"ContainerStarted","Data":"bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed"} Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.284983 4815 scope.go:117] "RemoveContainer" containerID="870d461dc7121d9c9d713dd5d5432e0d7aa0a0bfb6307ef895fbcc17b28d4c89" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.309747 4815 scope.go:117] "RemoveContainer" containerID="acb2fe51d5c916c6a38796623407e537767394a89e4b6186b451bc609db198de" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.325602 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rbpxh" podStartSLOduration=7.294442835 podStartE2EDuration="1m36.325582596s" podCreationTimestamp="2025-12-05 09:09:23 +0000 UTC" firstStartedPulling="2025-12-05 09:09:29.175582106 +0000 UTC m=+168.054188933" lastFinishedPulling="2025-12-05 09:10:58.206721857 +0000 UTC m=+257.085328694" observedRunningTime="2025-12-05 09:10:59.301092322 +0000 UTC m=+258.179699179" watchObservedRunningTime="2025-12-05 09:10:59.325582596 +0000 UTC m=+258.204189433" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.328144 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-whp68" podStartSLOduration=5.530742529 podStartE2EDuration="1m33.328135069s" podCreationTimestamp="2025-12-05 09:09:26 +0000 UTC" firstStartedPulling="2025-12-05 09:09:30.532190826 +0000 UTC m=+169.410797663" lastFinishedPulling="2025-12-05 09:10:58.329583366 +0000 UTC m=+257.208190203" observedRunningTime="2025-12-05 09:10:59.324187817 +0000 UTC m=+258.202794664" watchObservedRunningTime="2025-12-05 09:10:59.328135069 +0000 UTC m=+258.206741906" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.337956 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-q7gx9"] Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.341848 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-q7gx9"] Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.426990 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baadfde2-0beb-4d47-859e-bc80b697a7ea" path="/var/lib/kubelet/pods/baadfde2-0beb-4d47-859e-bc80b697a7ea/volumes" Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.656811 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jctlm"] Dec 05 09:10:59 crc kubenswrapper[4815]: I1205 09:10:59.657045 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jctlm" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerName="registry-server" containerID="cri-o://431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df" gracePeriod=2 Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.056971 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.150445 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-utilities\") pod \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.150913 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-catalog-content\") pod \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.151055 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2cgp\" (UniqueName: \"kubernetes.io/projected/d911f1b0-5e65-45ee-b2b6-d609bc40356c-kube-api-access-q2cgp\") pod \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\" (UID: \"d911f1b0-5e65-45ee-b2b6-d609bc40356c\") " Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.151415 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-utilities" (OuterVolumeSpecName: "utilities") pod "d911f1b0-5e65-45ee-b2b6-d609bc40356c" (UID: "d911f1b0-5e65-45ee-b2b6-d609bc40356c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.156429 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d911f1b0-5e65-45ee-b2b6-d609bc40356c-kube-api-access-q2cgp" (OuterVolumeSpecName: "kube-api-access-q2cgp") pod "d911f1b0-5e65-45ee-b2b6-d609bc40356c" (UID: "d911f1b0-5e65-45ee-b2b6-d609bc40356c"). InnerVolumeSpecName "kube-api-access-q2cgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.252397 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d911f1b0-5e65-45ee-b2b6-d609bc40356c" (UID: "d911f1b0-5e65-45ee-b2b6-d609bc40356c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.252509 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2cgp\" (UniqueName: \"kubernetes.io/projected/d911f1b0-5e65-45ee-b2b6-d609bc40356c-kube-api-access-q2cgp\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.252539 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.273284 4815 generic.go:334] "Generic (PLEG): container finished" podID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerID="431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df" exitCode=0 Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.273408 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jctlm" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.273833 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctlm" event={"ID":"d911f1b0-5e65-45ee-b2b6-d609bc40356c","Type":"ContainerDied","Data":"431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df"} Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.273866 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jctlm" event={"ID":"d911f1b0-5e65-45ee-b2b6-d609bc40356c","Type":"ContainerDied","Data":"6b8060aa79fafea63ad55104d4e478c64d6a322b1ce33d63c9c03c5aa044886c"} Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.273904 4815 scope.go:117] "RemoveContainer" containerID="431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.277348 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd8pw" event={"ID":"469b719a-7093-4818-ad40-87fa357a7e82","Type":"ContainerStarted","Data":"42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd"} Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.300615 4815 scope.go:117] "RemoveContainer" containerID="8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.320428 4815 scope.go:117] "RemoveContainer" containerID="43c495c9693386a162169551d421d4897c94bcd93464858f2281881723a7209b" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.322399 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rd8pw" podStartSLOduration=6.947160278 podStartE2EDuration="1m37.32236419s" podCreationTimestamp="2025-12-05 09:09:23 +0000 UTC" firstStartedPulling="2025-12-05 09:09:29.248366336 +0000 UTC m=+168.126973173" lastFinishedPulling="2025-12-05 09:10:59.623570248 +0000 UTC m=+258.502177085" observedRunningTime="2025-12-05 09:11:00.317421029 +0000 UTC m=+259.196027916" watchObservedRunningTime="2025-12-05 09:11:00.32236419 +0000 UTC m=+259.200971027" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.334898 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jctlm"] Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.337299 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jctlm"] Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.343097 4815 scope.go:117] "RemoveContainer" containerID="431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df" Dec 05 09:11:00 crc kubenswrapper[4815]: E1205 09:11:00.343396 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df\": container with ID starting with 431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df not found: ID does not exist" containerID="431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.343429 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df"} err="failed to get container status \"431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df\": rpc error: code = NotFound desc = could not find container \"431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df\": container with ID starting with 431553650b8fc9830842b86ab4fcfb64b8b6d5e8ab3276909363c6a254a724df not found: ID does not exist" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.343450 4815 scope.go:117] "RemoveContainer" containerID="8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7" Dec 05 09:11:00 crc kubenswrapper[4815]: E1205 09:11:00.343638 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7\": container with ID starting with 8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7 not found: ID does not exist" containerID="8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.343660 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7"} err="failed to get container status \"8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7\": rpc error: code = NotFound desc = could not find container \"8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7\": container with ID starting with 8cb6702f6e00cc678972848d4e7a9771d9abe62c62e29171a8b472c3cc2394e7 not found: ID does not exist" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.343672 4815 scope.go:117] "RemoveContainer" containerID="43c495c9693386a162169551d421d4897c94bcd93464858f2281881723a7209b" Dec 05 09:11:00 crc kubenswrapper[4815]: E1205 09:11:00.343816 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"43c495c9693386a162169551d421d4897c94bcd93464858f2281881723a7209b\": container with ID starting with 43c495c9693386a162169551d421d4897c94bcd93464858f2281881723a7209b not found: ID does not exist" containerID="43c495c9693386a162169551d421d4897c94bcd93464858f2281881723a7209b" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.343834 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"43c495c9693386a162169551d421d4897c94bcd93464858f2281881723a7209b"} err="failed to get container status \"43c495c9693386a162169551d421d4897c94bcd93464858f2281881723a7209b\": rpc error: code = NotFound desc = could not find container \"43c495c9693386a162169551d421d4897c94bcd93464858f2281881723a7209b\": container with ID starting with 43c495c9693386a162169551d421d4897c94bcd93464858f2281881723a7209b not found: ID does not exist" Dec 05 09:11:00 crc kubenswrapper[4815]: I1205 09:11:00.353639 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d911f1b0-5e65-45ee-b2b6-d609bc40356c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:01 crc kubenswrapper[4815]: I1205 09:11:01.428010 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" path="/var/lib/kubelet/pods/d911f1b0-5e65-45ee-b2b6-d609bc40356c/volumes" Dec 05 09:11:03 crc kubenswrapper[4815]: I1205 09:11:03.632605 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:11:03 crc kubenswrapper[4815]: I1205 09:11:03.632977 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:11:03 crc kubenswrapper[4815]: I1205 09:11:03.675248 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:11:04 crc kubenswrapper[4815]: I1205 09:11:04.358509 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:11:04 crc kubenswrapper[4815]: I1205 09:11:04.425886 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:11:04 crc kubenswrapper[4815]: I1205 09:11:04.426180 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:11:04 crc kubenswrapper[4815]: I1205 09:11:04.467795 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:11:04 crc kubenswrapper[4815]: I1205 09:11:04.467859 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:11:04 crc kubenswrapper[4815]: I1205 09:11:04.469860 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:11:04 crc kubenswrapper[4815]: I1205 09:11:04.519516 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:11:05 crc kubenswrapper[4815]: I1205 09:11:05.200472 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" podUID="a4d093b9-c855-48b0-9d44-c43fe89793bb" containerName="oauth-openshift" containerID="cri-o://2ab5e49b3488a45e7932a9f0d7452a3e9d80bf00677733341a04ef6b3456c083" gracePeriod=15 Dec 05 09:11:05 crc kubenswrapper[4815]: I1205 09:11:05.367334 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:11:05 crc kubenswrapper[4815]: I1205 09:11:05.384325 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:11:05 crc kubenswrapper[4815]: I1205 09:11:05.852640 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rd8pw"] Dec 05 09:11:06 crc kubenswrapper[4815]: I1205 09:11:06.876503 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:11:06 crc kubenswrapper[4815]: I1205 09:11:06.876821 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:11:06 crc kubenswrapper[4815]: I1205 09:11:06.932533 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.366776 4815 generic.go:334] "Generic (PLEG): container finished" podID="a4d093b9-c855-48b0-9d44-c43fe89793bb" containerID="2ab5e49b3488a45e7932a9f0d7452a3e9d80bf00677733341a04ef6b3456c083" exitCode=0 Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.366865 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" event={"ID":"a4d093b9-c855-48b0-9d44-c43fe89793bb","Type":"ContainerDied","Data":"2ab5e49b3488a45e7932a9f0d7452a3e9d80bf00677733341a04ef6b3456c083"} Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.367709 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rd8pw" podUID="469b719a-7093-4818-ad40-87fa357a7e82" containerName="registry-server" containerID="cri-o://42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd" gracePeriod=2 Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.412118 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.457174 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.501253 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-574dcf5686-rfnw4"] Dec 05 09:11:07 crc kubenswrapper[4815]: E1205 09:11:07.504126 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baadfde2-0beb-4d47-859e-bc80b697a7ea" containerName="extract-content" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504154 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="baadfde2-0beb-4d47-859e-bc80b697a7ea" containerName="extract-content" Dec 05 09:11:07 crc kubenswrapper[4815]: E1205 09:11:07.504166 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerName="extract-content" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504172 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerName="extract-content" Dec 05 09:11:07 crc kubenswrapper[4815]: E1205 09:11:07.504183 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6" containerName="pruner" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504190 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6" containerName="pruner" Dec 05 09:11:07 crc kubenswrapper[4815]: E1205 09:11:07.504197 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baadfde2-0beb-4d47-859e-bc80b697a7ea" containerName="extract-utilities" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504204 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="baadfde2-0beb-4d47-859e-bc80b697a7ea" containerName="extract-utilities" Dec 05 09:11:07 crc kubenswrapper[4815]: E1205 09:11:07.504214 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a4d093b9-c855-48b0-9d44-c43fe89793bb" containerName="oauth-openshift" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504220 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a4d093b9-c855-48b0-9d44-c43fe89793bb" containerName="oauth-openshift" Dec 05 09:11:07 crc kubenswrapper[4815]: E1205 09:11:07.504231 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerName="extract-utilities" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504236 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerName="extract-utilities" Dec 05 09:11:07 crc kubenswrapper[4815]: E1205 09:11:07.504245 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baadfde2-0beb-4d47-859e-bc80b697a7ea" containerName="registry-server" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504250 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="baadfde2-0beb-4d47-859e-bc80b697a7ea" containerName="registry-server" Dec 05 09:11:07 crc kubenswrapper[4815]: E1205 09:11:07.504262 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerName="registry-server" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504268 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerName="registry-server" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504386 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="baadfde2-0beb-4d47-859e-bc80b697a7ea" containerName="registry-server" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504397 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a4d093b9-c855-48b0-9d44-c43fe89793bb" containerName="oauth-openshift" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504409 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2af7ab6f-bc6a-4ed7-974b-c2bfb7f7efa6" containerName="pruner" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.504418 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d911f1b0-5e65-45ee-b2b6-d609bc40356c" containerName="registry-server" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.505705 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.535377 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-574dcf5686-rfnw4"] Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.552466 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-serving-cert\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.552555 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-service-ca\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.553386 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.553428 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2427\" (UniqueName: \"kubernetes.io/projected/a4d093b9-c855-48b0-9d44-c43fe89793bb-kube-api-access-x2427\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.553467 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-provider-selection\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.554039 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-router-certs\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.554065 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-trusted-ca-bundle\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.554101 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-login\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.554129 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-session\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.554149 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-cliconfig\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.554183 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-error\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.554205 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-ocp-branding-template\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.554235 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-policies\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.554257 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-dir\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.554289 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-idp-0-file-data\") pod \"a4d093b9-c855-48b0-9d44-c43fe89793bb\" (UID: \"a4d093b9-c855-48b0-9d44-c43fe89793bb\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.554537 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.555652 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.556782 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.557983 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.559282 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.559346 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.563744 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a4d093b9-c855-48b0-9d44-c43fe89793bb-kube-api-access-x2427" (OuterVolumeSpecName: "kube-api-access-x2427") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "kube-api-access-x2427". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.563907 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.566758 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.568055 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.579926 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.580685 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.580887 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.581081 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "a4d093b9-c855-48b0-9d44-c43fe89793bb" (UID: "a4d093b9-c855-48b0-9d44-c43fe89793bb"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656097 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-audit-policies\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656164 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-service-ca\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656216 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656267 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-template-login\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656289 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-template-error\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656412 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/90c29ff2-2ef3-4c5b-ba40-623d030ba101-audit-dir\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656540 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656567 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-serving-cert\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656609 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-session\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656629 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khwzc\" (UniqueName: \"kubernetes.io/projected/90c29ff2-2ef3-4c5b-ba40-623d030ba101-kube-api-access-khwzc\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656745 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-cliconfig\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656793 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656822 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-router-certs\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656866 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656953 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656966 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656977 4815 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656988 4815 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/a4d093b9-c855-48b0-9d44-c43fe89793bb-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.656998 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.657008 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.657018 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2427\" (UniqueName: \"kubernetes.io/projected/a4d093b9-c855-48b0-9d44-c43fe89793bb-kube-api-access-x2427\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.657027 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.657036 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.657044 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.657053 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.657062 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.657071 4815 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/a4d093b9-c855-48b0-9d44-c43fe89793bb-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.758733 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.758787 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-serving-cert\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.758820 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-session\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.758842 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khwzc\" (UniqueName: \"kubernetes.io/projected/90c29ff2-2ef3-4c5b-ba40-623d030ba101-kube-api-access-khwzc\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.758872 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-cliconfig\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.758896 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.758923 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-router-certs\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.758954 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.758988 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-audit-policies\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.759009 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-service-ca\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.759060 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.759090 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-template-login\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.759113 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-template-error\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.759145 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/90c29ff2-2ef3-4c5b-ba40-623d030ba101-audit-dir\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.759215 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/90c29ff2-2ef3-4c5b-ba40-623d030ba101-audit-dir\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.760319 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-cliconfig\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.760660 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.760723 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-audit-policies\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.761096 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-service-ca\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.764355 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-template-error\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.764532 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-serving-cert\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.764932 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-session\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.765615 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.766583 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-template-login\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.766641 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.767164 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.767846 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/90c29ff2-2ef3-4c5b-ba40-623d030ba101-v4-0-config-system-router-certs\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.781573 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khwzc\" (UniqueName: \"kubernetes.io/projected/90c29ff2-2ef3-4c5b-ba40-623d030ba101-kube-api-access-khwzc\") pod \"oauth-openshift-574dcf5686-rfnw4\" (UID: \"90c29ff2-2ef3-4c5b-ba40-623d030ba101\") " pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.785930 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.835181 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.961150 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5wrcp\" (UniqueName: \"kubernetes.io/projected/469b719a-7093-4818-ad40-87fa357a7e82-kube-api-access-5wrcp\") pod \"469b719a-7093-4818-ad40-87fa357a7e82\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.961203 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-catalog-content\") pod \"469b719a-7093-4818-ad40-87fa357a7e82\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.961244 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-utilities\") pod \"469b719a-7093-4818-ad40-87fa357a7e82\" (UID: \"469b719a-7093-4818-ad40-87fa357a7e82\") " Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.962240 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-utilities" (OuterVolumeSpecName: "utilities") pod "469b719a-7093-4818-ad40-87fa357a7e82" (UID: "469b719a-7093-4818-ad40-87fa357a7e82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:11:07 crc kubenswrapper[4815]: I1205 09:11:07.965801 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/469b719a-7093-4818-ad40-87fa357a7e82-kube-api-access-5wrcp" (OuterVolumeSpecName: "kube-api-access-5wrcp") pod "469b719a-7093-4818-ad40-87fa357a7e82" (UID: "469b719a-7093-4818-ad40-87fa357a7e82"). InnerVolumeSpecName "kube-api-access-5wrcp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.014703 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "469b719a-7093-4818-ad40-87fa357a7e82" (UID: "469b719a-7093-4818-ad40-87fa357a7e82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.062471 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5wrcp\" (UniqueName: \"kubernetes.io/projected/469b719a-7093-4818-ad40-87fa357a7e82-kube-api-access-5wrcp\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.062534 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.062552 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/469b719a-7093-4818-ad40-87fa357a7e82-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.235737 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-574dcf5686-rfnw4"] Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.373622 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.373613 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6sjd8" event={"ID":"a4d093b9-c855-48b0-9d44-c43fe89793bb","Type":"ContainerDied","Data":"981ced39b5b8476ddc5f4b4299f613f80107500064a1de0353b530ebc483e94b"} Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.373926 4815 scope.go:117] "RemoveContainer" containerID="2ab5e49b3488a45e7932a9f0d7452a3e9d80bf00677733341a04ef6b3456c083" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.374937 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" event={"ID":"90c29ff2-2ef3-4c5b-ba40-623d030ba101","Type":"ContainerStarted","Data":"caa71a6a0661899286657da9e254db58bbac72124466941dcafc0fd3e864ab4a"} Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.378166 4815 generic.go:334] "Generic (PLEG): container finished" podID="469b719a-7093-4818-ad40-87fa357a7e82" containerID="42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd" exitCode=0 Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.378224 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd8pw" event={"ID":"469b719a-7093-4818-ad40-87fa357a7e82","Type":"ContainerDied","Data":"42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd"} Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.378262 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rd8pw" event={"ID":"469b719a-7093-4818-ad40-87fa357a7e82","Type":"ContainerDied","Data":"74e70142247e705f9f996fe3a359ba8d3a175774e67ed52d807b4a7cddc373cd"} Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.378230 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rd8pw" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.393857 4815 scope.go:117] "RemoveContainer" containerID="42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.422017 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rd8pw"] Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.426706 4815 scope.go:117] "RemoveContainer" containerID="6edb26cdf605adc1b660738bb7f2ce1c8b98126cf1e3b6daeb99f1f0ca495a1c" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.444788 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rd8pw"] Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.448544 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjd8"] Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.449000 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjd8"] Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.481791 4815 scope.go:117] "RemoveContainer" containerID="e4baae428c8001a5e0d607fb2fc34d0edbfded41df22b20383d4bf4ea0fdb773" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.496528 4815 scope.go:117] "RemoveContainer" containerID="42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd" Dec 05 09:11:08 crc kubenswrapper[4815]: E1205 09:11:08.497174 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd\": container with ID starting with 42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd not found: ID does not exist" containerID="42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.497270 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd"} err="failed to get container status \"42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd\": rpc error: code = NotFound desc = could not find container \"42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd\": container with ID starting with 42d8b3bd8d132124cc8e4ee479d655f1636ac5657f4e17923ac265d8d98d59fd not found: ID does not exist" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.497352 4815 scope.go:117] "RemoveContainer" containerID="6edb26cdf605adc1b660738bb7f2ce1c8b98126cf1e3b6daeb99f1f0ca495a1c" Dec 05 09:11:08 crc kubenswrapper[4815]: E1205 09:11:08.497774 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6edb26cdf605adc1b660738bb7f2ce1c8b98126cf1e3b6daeb99f1f0ca495a1c\": container with ID starting with 6edb26cdf605adc1b660738bb7f2ce1c8b98126cf1e3b6daeb99f1f0ca495a1c not found: ID does not exist" containerID="6edb26cdf605adc1b660738bb7f2ce1c8b98126cf1e3b6daeb99f1f0ca495a1c" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.497852 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6edb26cdf605adc1b660738bb7f2ce1c8b98126cf1e3b6daeb99f1f0ca495a1c"} err="failed to get container status \"6edb26cdf605adc1b660738bb7f2ce1c8b98126cf1e3b6daeb99f1f0ca495a1c\": rpc error: code = NotFound desc = could not find container \"6edb26cdf605adc1b660738bb7f2ce1c8b98126cf1e3b6daeb99f1f0ca495a1c\": container with ID starting with 6edb26cdf605adc1b660738bb7f2ce1c8b98126cf1e3b6daeb99f1f0ca495a1c not found: ID does not exist" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.497932 4815 scope.go:117] "RemoveContainer" containerID="e4baae428c8001a5e0d607fb2fc34d0edbfded41df22b20383d4bf4ea0fdb773" Dec 05 09:11:08 crc kubenswrapper[4815]: E1205 09:11:08.498300 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4baae428c8001a5e0d607fb2fc34d0edbfded41df22b20383d4bf4ea0fdb773\": container with ID starting with e4baae428c8001a5e0d607fb2fc34d0edbfded41df22b20383d4bf4ea0fdb773 not found: ID does not exist" containerID="e4baae428c8001a5e0d607fb2fc34d0edbfded41df22b20383d4bf4ea0fdb773" Dec 05 09:11:08 crc kubenswrapper[4815]: I1205 09:11:08.498421 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4baae428c8001a5e0d607fb2fc34d0edbfded41df22b20383d4bf4ea0fdb773"} err="failed to get container status \"e4baae428c8001a5e0d607fb2fc34d0edbfded41df22b20383d4bf4ea0fdb773\": rpc error: code = NotFound desc = could not find container \"e4baae428c8001a5e0d607fb2fc34d0edbfded41df22b20383d4bf4ea0fdb773\": container with ID starting with e4baae428c8001a5e0d607fb2fc34d0edbfded41df22b20383d4bf4ea0fdb773 not found: ID does not exist" Dec 05 09:11:09 crc kubenswrapper[4815]: I1205 09:11:09.385362 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" event={"ID":"90c29ff2-2ef3-4c5b-ba40-623d030ba101","Type":"ContainerStarted","Data":"ede66a1819943a2a79e14b0bb4a14d465c5ec6d9e77d8f0b92a576697875277a"} Dec 05 09:11:09 crc kubenswrapper[4815]: I1205 09:11:09.386586 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:09 crc kubenswrapper[4815]: I1205 09:11:09.391073 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" Dec 05 09:11:09 crc kubenswrapper[4815]: I1205 09:11:09.403308 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-574dcf5686-rfnw4" podStartSLOduration=29.403293231 podStartE2EDuration="29.403293231s" podCreationTimestamp="2025-12-05 09:10:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:11:09.40255522 +0000 UTC m=+268.281162057" watchObservedRunningTime="2025-12-05 09:11:09.403293231 +0000 UTC m=+268.281900068" Dec 05 09:11:09 crc kubenswrapper[4815]: I1205 09:11:09.428059 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="469b719a-7093-4818-ad40-87fa357a7e82" path="/var/lib/kubelet/pods/469b719a-7093-4818-ad40-87fa357a7e82/volumes" Dec 05 09:11:09 crc kubenswrapper[4815]: I1205 09:11:09.428728 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a4d093b9-c855-48b0-9d44-c43fe89793bb" path="/var/lib/kubelet/pods/a4d093b9-c855-48b0-9d44-c43fe89793bb/volumes" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.234921 4815 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 09:11:10 crc kubenswrapper[4815]: E1205 09:11:10.235405 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469b719a-7093-4818-ad40-87fa357a7e82" containerName="extract-utilities" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.235507 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="469b719a-7093-4818-ad40-87fa357a7e82" containerName="extract-utilities" Dec 05 09:11:10 crc kubenswrapper[4815]: E1205 09:11:10.235602 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469b719a-7093-4818-ad40-87fa357a7e82" containerName="registry-server" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.235657 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="469b719a-7093-4818-ad40-87fa357a7e82" containerName="registry-server" Dec 05 09:11:10 crc kubenswrapper[4815]: E1205 09:11:10.235724 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="469b719a-7093-4818-ad40-87fa357a7e82" containerName="extract-content" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.235783 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="469b719a-7093-4818-ad40-87fa357a7e82" containerName="extract-content" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.235924 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="469b719a-7093-4818-ad40-87fa357a7e82" containerName="registry-server" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.236282 4815 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.236459 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.236692 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece" gracePeriod=15 Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.236796 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8" gracePeriod=15 Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.236804 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875" gracePeriod=15 Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.236834 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af" gracePeriod=15 Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.236796 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9" gracePeriod=15 Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.236919 4815 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 09:11:10 crc kubenswrapper[4815]: E1205 09:11:10.238650 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.238671 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 05 09:11:10 crc kubenswrapper[4815]: E1205 09:11:10.238685 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.238694 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 09:11:10 crc kubenswrapper[4815]: E1205 09:11:10.238706 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.238714 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 09:11:10 crc kubenswrapper[4815]: E1205 09:11:10.238724 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.238731 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 09:11:10 crc kubenswrapper[4815]: E1205 09:11:10.238746 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.238752 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 09:11:10 crc kubenswrapper[4815]: E1205 09:11:10.238774 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.238781 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 09:11:10 crc kubenswrapper[4815]: E1205 09:11:10.238792 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.238800 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.238965 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.238977 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.238990 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.239000 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.239013 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.239023 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.245294 4815 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.254106 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-whp68"] Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.254716 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-whp68" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" containerName="registry-server" containerID="cri-o://89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26" gracePeriod=2 Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.273734 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.390525 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.390578 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.390625 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.390666 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.390687 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.390715 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.390744 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.390763 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492268 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492365 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492393 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492441 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492561 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492586 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492623 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492663 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492818 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492851 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492851 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492932 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.492966 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.493018 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.493070 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.493140 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: I1205 09:11:10.566133 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:10 crc kubenswrapper[4815]: W1205 09:11:10.583627 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-88aa970ad5368852a53ba649fd0e2088c21bdfe0cf0228d26d5c4d89fd31e96d WatchSource:0}: Error finding container 88aa970ad5368852a53ba649fd0e2088c21bdfe0cf0228d26d5c4d89fd31e96d: Status 404 returned error can't find the container with id 88aa970ad5368852a53ba649fd0e2088c21bdfe0cf0228d26d5c4d89fd31e96d Dec 05 09:11:10 crc kubenswrapper[4815]: E1205 09:11:10.586519 4815 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.67:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e46ba732a86a7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 09:11:10.585841319 +0000 UTC m=+269.464448156,LastTimestamp:2025-12-05 09:11:10.585841319 +0000 UTC m=+269.464448156,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.175981 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.176658 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.177056 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.305380 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-utilities\") pod \"4d3dc703-3608-44cd-8993-58eaaef64058\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.305465 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-catalog-content\") pod \"4d3dc703-3608-44cd-8993-58eaaef64058\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.305566 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkmrq\" (UniqueName: \"kubernetes.io/projected/4d3dc703-3608-44cd-8993-58eaaef64058-kube-api-access-qkmrq\") pod \"4d3dc703-3608-44cd-8993-58eaaef64058\" (UID: \"4d3dc703-3608-44cd-8993-58eaaef64058\") " Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.306430 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-utilities" (OuterVolumeSpecName: "utilities") pod "4d3dc703-3608-44cd-8993-58eaaef64058" (UID: "4d3dc703-3608-44cd-8993-58eaaef64058"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.311056 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d3dc703-3608-44cd-8993-58eaaef64058-kube-api-access-qkmrq" (OuterVolumeSpecName: "kube-api-access-qkmrq") pod "4d3dc703-3608-44cd-8993-58eaaef64058" (UID: "4d3dc703-3608-44cd-8993-58eaaef64058"). InnerVolumeSpecName "kube-api-access-qkmrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.327210 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4d3dc703-3608-44cd-8993-58eaaef64058" (UID: "4d3dc703-3608-44cd-8993-58eaaef64058"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.400143 4815 generic.go:334] "Generic (PLEG): container finished" podID="2f986909-b717-447f-be83-09b47b6deb8d" containerID="9430f26de18c05a11a62c4a06d6a592ea91ac54193d51ef92383556784ab57ca" exitCode=0 Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.400213 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f986909-b717-447f-be83-09b47b6deb8d","Type":"ContainerDied","Data":"9430f26de18c05a11a62c4a06d6a592ea91ac54193d51ef92383556784ab57ca"} Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.400722 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.401010 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.401466 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.402137 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9"} Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.402173 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"88aa970ad5368852a53ba649fd0e2088c21bdfe0cf0228d26d5c4d89fd31e96d"} Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.403074 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.403391 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.403703 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.405395 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.406662 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.406718 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkmrq\" (UniqueName: \"kubernetes.io/projected/4d3dc703-3608-44cd-8993-58eaaef64058-kube-api-access-qkmrq\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.406743 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.406784 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4d3dc703-3608-44cd-8993-58eaaef64058-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.407320 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9" exitCode=0 Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.407349 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af" exitCode=0 Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.407360 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875" exitCode=0 Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.407372 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8" exitCode=2 Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.407410 4815 scope.go:117] "RemoveContainer" containerID="faa71436e789b680d8b51b24f0567d17eda24bc1952b1833c8b3b748b8dbba9c" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.409970 4815 generic.go:334] "Generic (PLEG): container finished" podID="4d3dc703-3608-44cd-8993-58eaaef64058" containerID="89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26" exitCode=0 Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.411056 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-whp68" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.412778 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whp68" event={"ID":"4d3dc703-3608-44cd-8993-58eaaef64058","Type":"ContainerDied","Data":"89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26"} Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.412832 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-whp68" event={"ID":"4d3dc703-3608-44cd-8993-58eaaef64058","Type":"ContainerDied","Data":"44fc1ff9db400c6f3f5a4712d10dd20317b886179c8959c347878497e5c16d00"} Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.413292 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.413466 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.413637 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.421695 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.422629 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.422866 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.443184 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.443918 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.444194 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.448811 4815 scope.go:117] "RemoveContainer" containerID="89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.469191 4815 scope.go:117] "RemoveContainer" containerID="e334c47c05dd792a9c646cbfa68e50b1e066559cf058b770d571df1d0e4a3931" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.491839 4815 scope.go:117] "RemoveContainer" containerID="63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.507779 4815 scope.go:117] "RemoveContainer" containerID="89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26" Dec 05 09:11:11 crc kubenswrapper[4815]: E1205 09:11:11.508443 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26\": container with ID starting with 89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26 not found: ID does not exist" containerID="89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.508476 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26"} err="failed to get container status \"89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26\": rpc error: code = NotFound desc = could not find container \"89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26\": container with ID starting with 89eddc3ae7e1814975fc8bba8fe7736710522599389247434b1865f09126fb26 not found: ID does not exist" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.508516 4815 scope.go:117] "RemoveContainer" containerID="e334c47c05dd792a9c646cbfa68e50b1e066559cf058b770d571df1d0e4a3931" Dec 05 09:11:11 crc kubenswrapper[4815]: E1205 09:11:11.508945 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e334c47c05dd792a9c646cbfa68e50b1e066559cf058b770d571df1d0e4a3931\": container with ID starting with e334c47c05dd792a9c646cbfa68e50b1e066559cf058b770d571df1d0e4a3931 not found: ID does not exist" containerID="e334c47c05dd792a9c646cbfa68e50b1e066559cf058b770d571df1d0e4a3931" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.508967 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e334c47c05dd792a9c646cbfa68e50b1e066559cf058b770d571df1d0e4a3931"} err="failed to get container status \"e334c47c05dd792a9c646cbfa68e50b1e066559cf058b770d571df1d0e4a3931\": rpc error: code = NotFound desc = could not find container \"e334c47c05dd792a9c646cbfa68e50b1e066559cf058b770d571df1d0e4a3931\": container with ID starting with e334c47c05dd792a9c646cbfa68e50b1e066559cf058b770d571df1d0e4a3931 not found: ID does not exist" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.508984 4815 scope.go:117] "RemoveContainer" containerID="63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca" Dec 05 09:11:11 crc kubenswrapper[4815]: E1205 09:11:11.509699 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca\": container with ID starting with 63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca not found: ID does not exist" containerID="63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca" Dec 05 09:11:11 crc kubenswrapper[4815]: I1205 09:11:11.509722 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca"} err="failed to get container status \"63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca\": rpc error: code = NotFound desc = could not find container \"63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca\": container with ID starting with 63a0be9f759872b0e7b3e3897536a37e46c40d352ed273010d371d586894adca not found: ID does not exist" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.420930 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.755533 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.757776 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.758272 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.758409 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.758447 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.758651 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.758877 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.759065 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.759239 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.759398 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.759566 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.763376 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-var-lock\") pod \"2f986909-b717-447f-be83-09b47b6deb8d\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.763408 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.763447 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f986909-b717-447f-be83-09b47b6deb8d-kube-api-access\") pod \"2f986909-b717-447f-be83-09b47b6deb8d\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.763473 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-kubelet-dir\") pod \"2f986909-b717-447f-be83-09b47b6deb8d\" (UID: \"2f986909-b717-447f-be83-09b47b6deb8d\") " Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.763496 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-var-lock" (OuterVolumeSpecName: "var-lock") pod "2f986909-b717-447f-be83-09b47b6deb8d" (UID: "2f986909-b717-447f-be83-09b47b6deb8d"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.763518 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.763562 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.763565 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.763617 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2f986909-b717-447f-be83-09b47b6deb8d" (UID: "2f986909-b717-447f-be83-09b47b6deb8d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.763644 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.763665 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.764003 4815 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.764024 4815 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.764034 4815 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-var-lock\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.764048 4815 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.764057 4815 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2f986909-b717-447f-be83-09b47b6deb8d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.768785 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f986909-b717-447f-be83-09b47b6deb8d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2f986909-b717-447f-be83-09b47b6deb8d" (UID: "2f986909-b717-447f-be83-09b47b6deb8d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:11:12 crc kubenswrapper[4815]: I1205 09:11:12.864452 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2f986909-b717-447f-be83-09b47b6deb8d-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.443165 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.443583 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.444956 4815 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece" exitCode=0 Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.445116 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.446131 4815 scope.go:117] "RemoveContainer" containerID="350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.446728 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.446802 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2f986909-b717-447f-be83-09b47b6deb8d","Type":"ContainerDied","Data":"1d96bf9e7b50203d2036f9620c3957e5b7c98ee0599eb6681ea4106dff24226d"} Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.446851 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d96bf9e7b50203d2036f9620c3957e5b7c98ee0599eb6681ea4106dff24226d" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.447098 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.447135 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.447452 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.455351 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.469471 4815 scope.go:117] "RemoveContainer" containerID="d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.484496 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.484880 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.485311 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.485950 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.486300 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.486546 4815 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.486763 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.487728 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.494728 4815 scope.go:117] "RemoveContainer" containerID="8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.508959 4815 scope.go:117] "RemoveContainer" containerID="6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.526737 4815 scope.go:117] "RemoveContainer" containerID="dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.544008 4815 scope.go:117] "RemoveContainer" containerID="117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.573122 4815 scope.go:117] "RemoveContainer" containerID="350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9" Dec 05 09:11:13 crc kubenswrapper[4815]: E1205 09:11:13.573945 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\": container with ID starting with 350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9 not found: ID does not exist" containerID="350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.573990 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9"} err="failed to get container status \"350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\": rpc error: code = NotFound desc = could not find container \"350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9\": container with ID starting with 350deb3cab4ffba778af00883626b59637284a44aadae546543022065bfc2ae9 not found: ID does not exist" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.574023 4815 scope.go:117] "RemoveContainer" containerID="d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af" Dec 05 09:11:13 crc kubenswrapper[4815]: E1205 09:11:13.574586 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\": container with ID starting with d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af not found: ID does not exist" containerID="d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.574631 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af"} err="failed to get container status \"d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\": rpc error: code = NotFound desc = could not find container \"d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af\": container with ID starting with d219582467a09f4c90725c9f74c242d11a2e1c86500ba72ee60b9c0e62c744af not found: ID does not exist" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.574665 4815 scope.go:117] "RemoveContainer" containerID="8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875" Dec 05 09:11:13 crc kubenswrapper[4815]: E1205 09:11:13.574930 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\": container with ID starting with 8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875 not found: ID does not exist" containerID="8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.574953 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875"} err="failed to get container status \"8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\": rpc error: code = NotFound desc = could not find container \"8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875\": container with ID starting with 8578e844264bd5d702496496d02a62838c188fb5b2676495032f1e1bac132875 not found: ID does not exist" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.574967 4815 scope.go:117] "RemoveContainer" containerID="6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8" Dec 05 09:11:13 crc kubenswrapper[4815]: E1205 09:11:13.575250 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\": container with ID starting with 6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8 not found: ID does not exist" containerID="6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.575276 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8"} err="failed to get container status \"6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\": rpc error: code = NotFound desc = could not find container \"6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8\": container with ID starting with 6a54e6a1d21bf5a7fbef5a9d997e01869d547a6417eaf00a55329854a0be23f8 not found: ID does not exist" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.575287 4815 scope.go:117] "RemoveContainer" containerID="dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece" Dec 05 09:11:13 crc kubenswrapper[4815]: E1205 09:11:13.575683 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\": container with ID starting with dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece not found: ID does not exist" containerID="dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.575751 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece"} err="failed to get container status \"dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\": rpc error: code = NotFound desc = could not find container \"dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece\": container with ID starting with dc6e97c23bcd47bcae137652486de49b40225f0c63474ef907ccc1a8d56faece not found: ID does not exist" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.575774 4815 scope.go:117] "RemoveContainer" containerID="117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86" Dec 05 09:11:13 crc kubenswrapper[4815]: E1205 09:11:13.576353 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\": container with ID starting with 117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86 not found: ID does not exist" containerID="117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86" Dec 05 09:11:13 crc kubenswrapper[4815]: I1205 09:11:13.576413 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86"} err="failed to get container status \"117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\": rpc error: code = NotFound desc = could not find container \"117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86\": container with ID starting with 117d37f0cdd263acbed26786cefa6114b66985fbfef29c78f9225b0bbfa05b86 not found: ID does not exist" Dec 05 09:11:15 crc kubenswrapper[4815]: E1205 09:11:15.074008 4815 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.129.56.67:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187e46ba732a86a7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-05 09:11:10.585841319 +0000 UTC m=+269.464448156,LastTimestamp:2025-12-05 09:11:10.585841319 +0000 UTC m=+269.464448156,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 05 09:11:15 crc kubenswrapper[4815]: E1205 09:11:15.308613 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:11:10Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:11:10Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:11:10Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-05T09:11:10Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:15adb3b2133604b064893f8009a74145e4c8bb5b134d111346dcccbdd2aa9bc2\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:164fc35a19aa6cc886c8015c8ee3eba4895e76b1152cb9d795e4f3154a8533a3\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1610512706},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:610b8d322265b2c9d6b07efb2be26bf4d91e428b46412d73f5bdae0218004794\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:eafb9c83c480396c34e85d1f5f5c2623be6305031245be36455850c0398bfcc7\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1209064267},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0029526507396e493c5dce1652c41ed9c239b29e84ee579a2735fdb1aa3bce83\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:e1d263cd2113e0727021ccf27c8a671f8cfeaefbf93d60e3a918d6f60c136c30\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201604946},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:e8990432556acad31519b1a73ec32f32d27c2034cf9e5cc4db8980efc7331594\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:ebe9f523f5c211a3a0f2570331dddcd5be15b12c1fecd9b8b121f881bfaad029\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1129027903},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:15 crc kubenswrapper[4815]: E1205 09:11:15.309353 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:15 crc kubenswrapper[4815]: E1205 09:11:15.309787 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:15 crc kubenswrapper[4815]: E1205 09:11:15.310097 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:15 crc kubenswrapper[4815]: E1205 09:11:15.310401 4815 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:15 crc kubenswrapper[4815]: E1205 09:11:15.310436 4815 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 05 09:11:15 crc kubenswrapper[4815]: E1205 09:11:15.499465 4815 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.129.56.67:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" volumeName="registry-storage" Dec 05 09:11:16 crc kubenswrapper[4815]: E1205 09:11:16.339224 4815 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:16 crc kubenswrapper[4815]: E1205 09:11:16.339864 4815 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:16 crc kubenswrapper[4815]: E1205 09:11:16.340094 4815 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:16 crc kubenswrapper[4815]: E1205 09:11:16.340408 4815 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:16 crc kubenswrapper[4815]: E1205 09:11:16.340893 4815 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:16 crc kubenswrapper[4815]: I1205 09:11:16.340976 4815 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 05 09:11:16 crc kubenswrapper[4815]: E1205 09:11:16.341328 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" interval="200ms" Dec 05 09:11:16 crc kubenswrapper[4815]: E1205 09:11:16.542729 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" interval="400ms" Dec 05 09:11:16 crc kubenswrapper[4815]: E1205 09:11:16.943934 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" interval="800ms" Dec 05 09:11:17 crc kubenswrapper[4815]: E1205 09:11:17.745408 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" interval="1.6s" Dec 05 09:11:19 crc kubenswrapper[4815]: E1205 09:11:19.346819 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" interval="3.2s" Dec 05 09:11:21 crc kubenswrapper[4815]: I1205 09:11:21.418076 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:21 crc kubenswrapper[4815]: I1205 09:11:21.422345 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:21 crc kubenswrapper[4815]: I1205 09:11:21.422733 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:21 crc kubenswrapper[4815]: I1205 09:11:21.423070 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:21 crc kubenswrapper[4815]: I1205 09:11:21.423330 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:21 crc kubenswrapper[4815]: I1205 09:11:21.423540 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:21 crc kubenswrapper[4815]: I1205 09:11:21.423727 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:21 crc kubenswrapper[4815]: I1205 09:11:21.440583 4815 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5e44bbe8-6099-453d-8d03-275929053b58" Dec 05 09:11:21 crc kubenswrapper[4815]: I1205 09:11:21.440636 4815 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5e44bbe8-6099-453d-8d03-275929053b58" Dec 05 09:11:21 crc kubenswrapper[4815]: E1205 09:11:21.441243 4815 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:21 crc kubenswrapper[4815]: I1205 09:11:21.446962 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:21 crc kubenswrapper[4815]: W1205 09:11:21.470506 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-62a2bbdffb6267385717a4a03e19944cd6b52a274b477b272b96f701bec7e6e1 WatchSource:0}: Error finding container 62a2bbdffb6267385717a4a03e19944cd6b52a274b477b272b96f701bec7e6e1: Status 404 returned error can't find the container with id 62a2bbdffb6267385717a4a03e19944cd6b52a274b477b272b96f701bec7e6e1 Dec 05 09:11:21 crc kubenswrapper[4815]: I1205 09:11:21.494360 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"62a2bbdffb6267385717a4a03e19944cd6b52a274b477b272b96f701bec7e6e1"} Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.499403 4815 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="b6b13b90c9347c8ffaa1ec4ef59a196ce3eef143c23170fefb7899ac7752790b" exitCode=0 Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.499720 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"b6b13b90c9347c8ffaa1ec4ef59a196ce3eef143c23170fefb7899ac7752790b"} Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.500161 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.500388 4815 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5e44bbe8-6099-453d-8d03-275929053b58" Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.500497 4815 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5e44bbe8-6099-453d-8d03-275929053b58" Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.500393 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.500874 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:22 crc kubenswrapper[4815]: E1205 09:11:22.501002 4815 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.502044 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.502072 4815 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="3a037ee90341ab082b8523235fd47661c15400baae0eaaa3f54340b6f5885962" exitCode=1 Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.502089 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"3a037ee90341ab082b8523235fd47661c15400baae0eaaa3f54340b6f5885962"} Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.502595 4815 scope.go:117] "RemoveContainer" containerID="3a037ee90341ab082b8523235fd47661c15400baae0eaaa3f54340b6f5885962" Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.503244 4815 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.503422 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.503606 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:22 crc kubenswrapper[4815]: I1205 09:11:22.503778 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:22 crc kubenswrapper[4815]: E1205 09:11:22.548852 4815 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.129.56.67:6443: connect: connection refused" interval="6.4s" Dec 05 09:11:23 crc kubenswrapper[4815]: I1205 09:11:23.511630 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 05 09:11:23 crc kubenswrapper[4815]: I1205 09:11:23.511697 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6f3383398388a81a3821e1c503af6e4d3caac003e8cca32552c66798d941242d"} Dec 05 09:11:23 crc kubenswrapper[4815]: I1205 09:11:23.512795 4815 status_manager.go:851] "Failed to get status for pod" podUID="2f986909-b717-447f-be83-09b47b6deb8d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:23 crc kubenswrapper[4815]: I1205 09:11:23.512940 4815 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:23 crc kubenswrapper[4815]: I1205 09:11:23.513075 4815 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:23 crc kubenswrapper[4815]: I1205 09:11:23.513218 4815 status_manager.go:851] "Failed to get status for pod" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" pod="openshift-marketplace/redhat-marketplace-whp68" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-whp68\": dial tcp 38.129.56.67:6443: connect: connection refused" Dec 05 09:11:23 crc kubenswrapper[4815]: I1205 09:11:23.517326 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b02f958cf88898667b807b1b41b530592dcdc13b7ce23ec734d452a71a701f8c"} Dec 05 09:11:24 crc kubenswrapper[4815]: I1205 09:11:24.546296 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d776158e924cbb90019c02246fc9eaeb5911b40b14924044a4719043e4d3d92b"} Dec 05 09:11:24 crc kubenswrapper[4815]: I1205 09:11:24.546507 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"026597c9e7f0309394a4bdfcdd0ef8ae84fdd6cf372f545926f08f6d2bb531bc"} Dec 05 09:11:24 crc kubenswrapper[4815]: I1205 09:11:24.546520 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"19f2b40eaba777deb64ff15b0212284b556d618491942c1e94a3bd4a1eea05fd"} Dec 05 09:11:24 crc kubenswrapper[4815]: I1205 09:11:24.546530 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"74fe38977d4446ce9882bfff995761c3cc70053bd705b6b7594b93249a578293"} Dec 05 09:11:24 crc kubenswrapper[4815]: I1205 09:11:24.546749 4815 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5e44bbe8-6099-453d-8d03-275929053b58" Dec 05 09:11:24 crc kubenswrapper[4815]: I1205 09:11:24.546763 4815 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5e44bbe8-6099-453d-8d03-275929053b58" Dec 05 09:11:24 crc kubenswrapper[4815]: I1205 09:11:24.547101 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:26 crc kubenswrapper[4815]: I1205 09:11:26.447593 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:26 crc kubenswrapper[4815]: I1205 09:11:26.447894 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:26 crc kubenswrapper[4815]: I1205 09:11:26.452831 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:26 crc kubenswrapper[4815]: I1205 09:11:26.909638 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:11:26 crc kubenswrapper[4815]: I1205 09:11:26.910327 4815 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 05 09:11:26 crc kubenswrapper[4815]: I1205 09:11:26.910428 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 05 09:11:28 crc kubenswrapper[4815]: I1205 09:11:28.996125 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:11:30 crc kubenswrapper[4815]: I1205 09:11:30.062392 4815 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:30 crc kubenswrapper[4815]: I1205 09:11:30.576066 4815 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5e44bbe8-6099-453d-8d03-275929053b58" Dec 05 09:11:30 crc kubenswrapper[4815]: I1205 09:11:30.576094 4815 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5e44bbe8-6099-453d-8d03-275929053b58" Dec 05 09:11:30 crc kubenswrapper[4815]: I1205 09:11:30.583810 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:31 crc kubenswrapper[4815]: I1205 09:11:31.452780 4815 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7adce437-b99a-40c0-b4ad-60e38a6a047d" Dec 05 09:11:31 crc kubenswrapper[4815]: I1205 09:11:31.580522 4815 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5e44bbe8-6099-453d-8d03-275929053b58" Dec 05 09:11:31 crc kubenswrapper[4815]: I1205 09:11:31.580554 4815 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="5e44bbe8-6099-453d-8d03-275929053b58" Dec 05 09:11:31 crc kubenswrapper[4815]: I1205 09:11:31.583242 4815 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="7adce437-b99a-40c0-b4ad-60e38a6a047d" Dec 05 09:11:36 crc kubenswrapper[4815]: I1205 09:11:36.909590 4815 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Dec 05 09:11:36 crc kubenswrapper[4815]: I1205 09:11:36.909648 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Dec 05 09:11:41 crc kubenswrapper[4815]: I1205 09:11:41.065677 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 09:11:41 crc kubenswrapper[4815]: I1205 09:11:41.215287 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 05 09:11:41 crc kubenswrapper[4815]: I1205 09:11:41.274437 4815 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Dec 05 09:11:41 crc kubenswrapper[4815]: I1205 09:11:41.770083 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 05 09:11:42 crc kubenswrapper[4815]: I1205 09:11:42.050462 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 05 09:11:42 crc kubenswrapper[4815]: I1205 09:11:42.251923 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 05 09:11:42 crc kubenswrapper[4815]: I1205 09:11:42.313921 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 05 09:11:42 crc kubenswrapper[4815]: I1205 09:11:42.356771 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 05 09:11:42 crc kubenswrapper[4815]: I1205 09:11:42.584193 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 05 09:11:42 crc kubenswrapper[4815]: I1205 09:11:42.758116 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 05 09:11:42 crc kubenswrapper[4815]: I1205 09:11:42.763077 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 05 09:11:42 crc kubenswrapper[4815]: I1205 09:11:42.832797 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 05 09:11:42 crc kubenswrapper[4815]: I1205 09:11:42.844859 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.039808 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.046328 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.076326 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.234931 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.329233 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.379050 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.387302 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.533556 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.566845 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.605014 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.616883 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.685212 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.746993 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.801251 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.841974 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.844130 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 05 09:11:43 crc kubenswrapper[4815]: I1205 09:11:43.916176 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.023953 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.037087 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.200001 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.239877 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.293482 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.326933 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.386452 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.388049 4815 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.436630 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.504165 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.508990 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.519272 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.606195 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.607596 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.645297 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.698440 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.766641 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.855106 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.870989 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.879600 4815 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.886394 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 09:11:44 crc kubenswrapper[4815]: I1205 09:11:44.913541 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.047671 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.086403 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.129689 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.167462 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.218147 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.241337 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.346831 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.356185 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.471239 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.526886 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.600428 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.640178 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.825363 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.839818 4815 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.914096 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.941007 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.941291 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 05 09:11:45 crc kubenswrapper[4815]: I1205 09:11:45.995405 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.109151 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.152323 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.200057 4815 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.203310 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=36.203294405 podStartE2EDuration="36.203294405s" podCreationTimestamp="2025-12-05 09:11:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:11:30.076279789 +0000 UTC m=+288.954886626" watchObservedRunningTime="2025-12-05 09:11:46.203294405 +0000 UTC m=+305.081901242" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.204313 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-whp68","openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.204359 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.210146 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.224451 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.224428635 podStartE2EDuration="16.224428635s" podCreationTimestamp="2025-12-05 09:11:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:11:46.220935255 +0000 UTC m=+305.099542102" watchObservedRunningTime="2025-12-05 09:11:46.224428635 +0000 UTC m=+305.103035482" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.260864 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.273146 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.277236 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.285906 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.321887 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.325631 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.352233 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.416404 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.481397 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.540103 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.594022 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.614143 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.645063 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.722289 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.815671 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.913561 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:11:46 crc kubenswrapper[4815]: I1205 09:11:46.918496 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.163904 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.176483 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.176513 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.203010 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.207067 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.217163 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.239437 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.424876 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" path="/var/lib/kubelet/pods/4d3dc703-3608-44cd-8993-58eaaef64058/volumes" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.639392 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.655967 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.705062 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.746071 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.767658 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.829640 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 05 09:11:47 crc kubenswrapper[4815]: I1205 09:11:47.878924 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.060913 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.292514 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.331749 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.379076 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.386014 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.405314 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.436710 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.490331 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.497860 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.522787 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.549690 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.585879 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.595618 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.610572 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.685385 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.704624 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.757854 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.806813 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.832924 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.882540 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 05 09:11:48 crc kubenswrapper[4815]: I1205 09:11:48.887848 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.001336 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.035356 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.095422 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.187673 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.373434 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.425947 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.426219 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.426478 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.428270 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.457369 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.460216 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.503746 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.710572 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.724204 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.731885 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.807309 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.888951 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 05 09:11:49 crc kubenswrapper[4815]: I1205 09:11:49.950426 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.050926 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.051439 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.062510 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.064272 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.096869 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.137766 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.202331 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.261617 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.284332 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.346267 4815 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.374186 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.495917 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.509150 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.543440 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.637336 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.638036 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.640924 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.677581 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.794604 4815 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.818961 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.839520 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.881829 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.982783 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.983420 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 05 09:11:50 crc kubenswrapper[4815]: I1205 09:11:50.993795 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.018207 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.038455 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.048240 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.185794 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.305083 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.445958 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.476000 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.486545 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.502892 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.514687 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.564156 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.605377 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.683659 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.797147 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.842848 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 05 09:11:51 crc kubenswrapper[4815]: I1205 09:11:51.888658 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.055819 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.224858 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.249774 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.257983 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.307828 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.346562 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.465251 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.690636 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.695859 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.743681 4815 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.743931 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9" gracePeriod=5 Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.786127 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.803750 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 05 09:11:52 crc kubenswrapper[4815]: I1205 09:11:52.926176 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.018342 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.033645 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.173041 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.173920 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.250598 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.251247 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.348683 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.401676 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.433047 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.605608 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.679831 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.782870 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.838004 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 09:11:53 crc kubenswrapper[4815]: I1205 09:11:53.968697 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.077837 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.088589 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.093606 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.120100 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.181588 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.418107 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.454786 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.460698 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.496574 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.561291 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.565003 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.619691 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.639438 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.790024 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.894153 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.907794 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 05 09:11:54 crc kubenswrapper[4815]: I1205 09:11:54.974010 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.019874 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.027306 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.032057 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.051659 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.110074 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.156799 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.296207 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.376890 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.440274 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.561289 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.660382 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.813423 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.830873 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 05 09:11:55 crc kubenswrapper[4815]: I1205 09:11:55.915518 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 05 09:11:56 crc kubenswrapper[4815]: I1205 09:11:56.194418 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 09:11:56 crc kubenswrapper[4815]: I1205 09:11:56.212542 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 05 09:11:56 crc kubenswrapper[4815]: I1205 09:11:56.648218 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 05 09:11:56 crc kubenswrapper[4815]: I1205 09:11:56.695965 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 05 09:11:56 crc kubenswrapper[4815]: I1205 09:11:56.790539 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 05 09:11:56 crc kubenswrapper[4815]: I1205 09:11:56.853192 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 05 09:11:56 crc kubenswrapper[4815]: I1205 09:11:56.942588 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 05 09:11:57 crc kubenswrapper[4815]: I1205 09:11:57.062571 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 05 09:11:57 crc kubenswrapper[4815]: I1205 09:11:57.064137 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 05 09:11:57 crc kubenswrapper[4815]: I1205 09:11:57.246992 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 05 09:11:57 crc kubenswrapper[4815]: I1205 09:11:57.401056 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 05 09:11:57 crc kubenswrapper[4815]: I1205 09:11:57.712174 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.466699 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.466960 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.644042 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.644864 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.644902 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.644957 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.644981 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.645113 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.645167 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.645186 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.645207 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.652737 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.719914 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.719965 4815 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9" exitCode=137 Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.720009 4815 scope.go:117] "RemoveContainer" containerID="d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.720117 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.746080 4815 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.746218 4815 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.746285 4815 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.746349 4815 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.746407 4815 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.751521 4815 scope.go:117] "RemoveContainer" containerID="d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9" Dec 05 09:11:58 crc kubenswrapper[4815]: E1205 09:11:58.752073 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9\": container with ID starting with d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9 not found: ID does not exist" containerID="d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.752177 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9"} err="failed to get container status \"d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9\": rpc error: code = NotFound desc = could not find container \"d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9\": container with ID starting with d464382ae4886a2871b47b126cb5548ccfb8cec2309de937661e7dd1ec33ccf9 not found: ID does not exist" Dec 05 09:11:58 crc kubenswrapper[4815]: I1205 09:11:58.967954 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 05 09:11:59 crc kubenswrapper[4815]: I1205 09:11:59.237877 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 05 09:11:59 crc kubenswrapper[4815]: I1205 09:11:59.424394 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 05 09:11:59 crc kubenswrapper[4815]: I1205 09:11:59.424959 4815 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Dec 05 09:11:59 crc kubenswrapper[4815]: I1205 09:11:59.435880 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 09:11:59 crc kubenswrapper[4815]: I1205 09:11:59.436086 4815 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="3dcaa185-7fa5-47a6-9898-abcaecd44a01" Dec 05 09:11:59 crc kubenswrapper[4815]: I1205 09:11:59.439019 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 05 09:11:59 crc kubenswrapper[4815]: I1205 09:11:59.439056 4815 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="3dcaa185-7fa5-47a6-9898-abcaecd44a01" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.090914 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rbpxh"] Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.091160 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rbpxh" podUID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" containerName="registry-server" containerID="cri-o://bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed" gracePeriod=30 Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.097603 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nntc6"] Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.098157 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nntc6" podUID="201dec63-11da-4e0d-93ae-025de8a76beb" containerName="registry-server" containerID="cri-o://a5974d90794dd4068b02899a36ee5661556df8ec420b6f0ea92704a153f3018d" gracePeriod=30 Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.107994 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgk4c"] Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.108236 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" podUID="15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" containerName="marketplace-operator" containerID="cri-o://83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d" gracePeriod=30 Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.120312 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7nc7"] Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.120615 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-k7nc7" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerName="registry-server" containerID="cri-o://e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f" gracePeriod=30 Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.140580 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kb5lk"] Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.140958 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-kb5lk" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerName="registry-server" containerID="cri-o://236ca6b864b6e63b6a57d4677aee4d7e6c8fe6116c9aede3d6af76b45a3ec9d8" gracePeriod=30 Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.669267 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.673587 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.699764 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.742703 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.745170 4815 generic.go:334] "Generic (PLEG): container finished" podID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" containerID="bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed" exitCode=0 Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.745259 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbpxh" event={"ID":"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d","Type":"ContainerDied","Data":"bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed"} Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.745533 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rbpxh" event={"ID":"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d","Type":"ContainerDied","Data":"8177ec92aa411db5db7dda89a7ed4591ea1276fde7ae597d6ca37902e560ac68"} Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.745560 4815 scope.go:117] "RemoveContainer" containerID="bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.745925 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rbpxh" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.748473 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.752021 4815 generic.go:334] "Generic (PLEG): container finished" podID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerID="236ca6b864b6e63b6a57d4677aee4d7e6c8fe6116c9aede3d6af76b45a3ec9d8" exitCode=0 Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.752086 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-kb5lk" event={"ID":"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a","Type":"ContainerDied","Data":"236ca6b864b6e63b6a57d4677aee4d7e6c8fe6116c9aede3d6af76b45a3ec9d8"} Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.752504 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-kb5lk" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.767569 4815 generic.go:334] "Generic (PLEG): container finished" podID="201dec63-11da-4e0d-93ae-025de8a76beb" containerID="a5974d90794dd4068b02899a36ee5661556df8ec420b6f0ea92704a153f3018d" exitCode=0 Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.767641 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nntc6" event={"ID":"201dec63-11da-4e0d-93ae-025de8a76beb","Type":"ContainerDied","Data":"a5974d90794dd4068b02899a36ee5661556df8ec420b6f0ea92704a153f3018d"} Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.767719 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nntc6" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.772801 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-utilities\") pod \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.772847 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jj8r\" (UniqueName: \"kubernetes.io/projected/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-kube-api-access-4jj8r\") pod \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.772869 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-trusted-ca\") pod \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.772891 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-catalog-content\") pod \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.772920 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljzgn\" (UniqueName: \"kubernetes.io/projected/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-kube-api-access-ljzgn\") pod \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\" (UID: \"8084d369-d4ee-4dad-bdd3-c8d7ea34199c\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.772979 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-operator-metrics\") pod \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\" (UID: \"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.774434 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-utilities" (OuterVolumeSpecName: "utilities") pod "8084d369-d4ee-4dad-bdd3-c8d7ea34199c" (UID: "8084d369-d4ee-4dad-bdd3-c8d7ea34199c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.775134 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" (UID: "15a8ad31-81dd-4764-b4da-3c9b3eaa3b25"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.780933 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" (UID: "15a8ad31-81dd-4764-b4da-3c9b3eaa3b25"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.780953 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-kube-api-access-ljzgn" (OuterVolumeSpecName: "kube-api-access-ljzgn") pod "8084d369-d4ee-4dad-bdd3-c8d7ea34199c" (UID: "8084d369-d4ee-4dad-bdd3-c8d7ea34199c"). InnerVolumeSpecName "kube-api-access-ljzgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.793145 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-k7nc7" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.794010 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7nc7" event={"ID":"8084d369-d4ee-4dad-bdd3-c8d7ea34199c","Type":"ContainerDied","Data":"e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f"} Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.795574 4815 generic.go:334] "Generic (PLEG): container finished" podID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerID="e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f" exitCode=0 Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.795861 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-k7nc7" event={"ID":"8084d369-d4ee-4dad-bdd3-c8d7ea34199c","Type":"ContainerDied","Data":"752903add4dd69f6eaad5077224a283159d390620ec9a5714742e10e027269af"} Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.797648 4815 scope.go:117] "RemoveContainer" containerID="5ea687eb75a04e26956a84004a193937e60aa9219245d462e6aa8d026d548cdb" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.801184 4815 generic.go:334] "Generic (PLEG): container finished" podID="15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" containerID="83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d" exitCode=0 Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.801221 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" event={"ID":"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25","Type":"ContainerDied","Data":"83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d"} Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.801245 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" event={"ID":"15a8ad31-81dd-4764-b4da-3c9b3eaa3b25","Type":"ContainerDied","Data":"51b0fdaca01f9eb47423f3292fbee8b69e62ba341caef35f45bc55c37833aed9"} Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.801429 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-dgk4c" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.805785 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8084d369-d4ee-4dad-bdd3-c8d7ea34199c" (UID: "8084d369-d4ee-4dad-bdd3-c8d7ea34199c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.815332 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-kube-api-access-4jj8r" (OuterVolumeSpecName: "kube-api-access-4jj8r") pod "15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" (UID: "15a8ad31-81dd-4764-b4da-3c9b3eaa3b25"). InnerVolumeSpecName "kube-api-access-4jj8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.826776 4815 scope.go:117] "RemoveContainer" containerID="dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.851733 4815 scope.go:117] "RemoveContainer" containerID="bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed" Dec 05 09:12:00 crc kubenswrapper[4815]: E1205 09:12:00.852302 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed\": container with ID starting with bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed not found: ID does not exist" containerID="bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.852352 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed"} err="failed to get container status \"bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed\": rpc error: code = NotFound desc = could not find container \"bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed\": container with ID starting with bfaca117c0a90f1143c351a0f68abbcbe67b34d88bd7c31240cbfbc53e5944ed not found: ID does not exist" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.852379 4815 scope.go:117] "RemoveContainer" containerID="5ea687eb75a04e26956a84004a193937e60aa9219245d462e6aa8d026d548cdb" Dec 05 09:12:00 crc kubenswrapper[4815]: E1205 09:12:00.855140 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ea687eb75a04e26956a84004a193937e60aa9219245d462e6aa8d026d548cdb\": container with ID starting with 5ea687eb75a04e26956a84004a193937e60aa9219245d462e6aa8d026d548cdb not found: ID does not exist" containerID="5ea687eb75a04e26956a84004a193937e60aa9219245d462e6aa8d026d548cdb" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.855173 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ea687eb75a04e26956a84004a193937e60aa9219245d462e6aa8d026d548cdb"} err="failed to get container status \"5ea687eb75a04e26956a84004a193937e60aa9219245d462e6aa8d026d548cdb\": rpc error: code = NotFound desc = could not find container \"5ea687eb75a04e26956a84004a193937e60aa9219245d462e6aa8d026d548cdb\": container with ID starting with 5ea687eb75a04e26956a84004a193937e60aa9219245d462e6aa8d026d548cdb not found: ID does not exist" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.855190 4815 scope.go:117] "RemoveContainer" containerID="dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1" Dec 05 09:12:00 crc kubenswrapper[4815]: E1205 09:12:00.855462 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1\": container with ID starting with dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1 not found: ID does not exist" containerID="dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.855529 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1"} err="failed to get container status \"dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1\": rpc error: code = NotFound desc = could not find container \"dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1\": container with ID starting with dc0feaf4eb85b81448c764887a6185e83a7d187033f37a3dd6a6371203eee1c1 not found: ID does not exist" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.855566 4815 scope.go:117] "RemoveContainer" containerID="236ca6b864b6e63b6a57d4677aee4d7e6c8fe6116c9aede3d6af76b45a3ec9d8" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.870570 4815 scope.go:117] "RemoveContainer" containerID="7b3969eeb2c2ac590d775c30bacd72f6a8027ecae7e46b4a437aeb1743b7adda" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.873671 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7hkkk\" (UniqueName: \"kubernetes.io/projected/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-kube-api-access-7hkkk\") pod \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.873713 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-utilities\") pod \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.873756 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-utilities\") pod \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.873772 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-catalog-content\") pod \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.873800 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-catalog-content\") pod \"201dec63-11da-4e0d-93ae-025de8a76beb\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.873818 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l4nfq\" (UniqueName: \"kubernetes.io/projected/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-kube-api-access-l4nfq\") pod \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\" (UID: \"e7fa3908-d40f-43d1-9f59-7ad5c8a9877a\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.873835 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-catalog-content\") pod \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\" (UID: \"aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.873856 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-utilities\") pod \"201dec63-11da-4e0d-93ae-025de8a76beb\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.873876 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l6ct\" (UniqueName: \"kubernetes.io/projected/201dec63-11da-4e0d-93ae-025de8a76beb-kube-api-access-8l6ct\") pod \"201dec63-11da-4e0d-93ae-025de8a76beb\" (UID: \"201dec63-11da-4e0d-93ae-025de8a76beb\") " Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.874041 4815 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.874054 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.874063 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jj8r\" (UniqueName: \"kubernetes.io/projected/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-kube-api-access-4jj8r\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.874071 4815 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.874079 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.874089 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljzgn\" (UniqueName: \"kubernetes.io/projected/8084d369-d4ee-4dad-bdd3-c8d7ea34199c-kube-api-access-ljzgn\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.876654 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-utilities" (OuterVolumeSpecName: "utilities") pod "e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" (UID: "e7fa3908-d40f-43d1-9f59-7ad5c8a9877a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.876850 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-utilities" (OuterVolumeSpecName: "utilities") pod "201dec63-11da-4e0d-93ae-025de8a76beb" (UID: "201dec63-11da-4e0d-93ae-025de8a76beb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.877549 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/201dec63-11da-4e0d-93ae-025de8a76beb-kube-api-access-8l6ct" (OuterVolumeSpecName: "kube-api-access-8l6ct") pod "201dec63-11da-4e0d-93ae-025de8a76beb" (UID: "201dec63-11da-4e0d-93ae-025de8a76beb"). InnerVolumeSpecName "kube-api-access-8l6ct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.877806 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-utilities" (OuterVolumeSpecName: "utilities") pod "aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" (UID: "aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.878924 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-kube-api-access-7hkkk" (OuterVolumeSpecName: "kube-api-access-7hkkk") pod "aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" (UID: "aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d"). InnerVolumeSpecName "kube-api-access-7hkkk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.879199 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-kube-api-access-l4nfq" (OuterVolumeSpecName: "kube-api-access-l4nfq") pod "e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" (UID: "e7fa3908-d40f-43d1-9f59-7ad5c8a9877a"). InnerVolumeSpecName "kube-api-access-l4nfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.897417 4815 scope.go:117] "RemoveContainer" containerID="bf07c94210d582f53b847a2eeaf053d2f5cda30b8f655128e29b65ce9a9ac371" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.927777 4815 scope.go:117] "RemoveContainer" containerID="a5974d90794dd4068b02899a36ee5661556df8ec420b6f0ea92704a153f3018d" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.928618 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" (UID: "aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.950693 4815 scope.go:117] "RemoveContainer" containerID="f5d15f79a48da0117d5b75e7d92e9451a6d191594f117b1d95195379fc6587f2" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.964187 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "201dec63-11da-4e0d-93ae-025de8a76beb" (UID: "201dec63-11da-4e0d-93ae-025de8a76beb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.966754 4815 scope.go:117] "RemoveContainer" containerID="c2e07e58596989263ae33bf893e9506ca5e7fbc0089de6723765ca74e627a5cb" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.975380 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7hkkk\" (UniqueName: \"kubernetes.io/projected/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-kube-api-access-7hkkk\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.975402 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.975412 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.975422 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.975430 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l4nfq\" (UniqueName: \"kubernetes.io/projected/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-kube-api-access-l4nfq\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.975439 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/201dec63-11da-4e0d-93ae-025de8a76beb-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.975447 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.975455 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l6ct\" (UniqueName: \"kubernetes.io/projected/201dec63-11da-4e0d-93ae-025de8a76beb-kube-api-access-8l6ct\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.979099 4815 scope.go:117] "RemoveContainer" containerID="e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f" Dec 05 09:12:00 crc kubenswrapper[4815]: I1205 09:12:00.989192 4815 scope.go:117] "RemoveContainer" containerID="a04066155246b1f7e94c0699b459498b72601def19cc4a9ffeabaf9ea91b5017" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.001629 4815 scope.go:117] "RemoveContainer" containerID="52d80bc9c6229451f2e24f36468de30d9ed19786d83228e3b2cbe9b82c63b7ce" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.014122 4815 scope.go:117] "RemoveContainer" containerID="e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f" Dec 05 09:12:01 crc kubenswrapper[4815]: E1205 09:12:01.014502 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f\": container with ID starting with e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f not found: ID does not exist" containerID="e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.014543 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f"} err="failed to get container status \"e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f\": rpc error: code = NotFound desc = could not find container \"e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f\": container with ID starting with e09e8c7abc5c8284543f7ee064b1cf991b0a9746f4c609f04fd5aa7f6f9e9b1f not found: ID does not exist" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.014571 4815 scope.go:117] "RemoveContainer" containerID="a04066155246b1f7e94c0699b459498b72601def19cc4a9ffeabaf9ea91b5017" Dec 05 09:12:01 crc kubenswrapper[4815]: E1205 09:12:01.014903 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a04066155246b1f7e94c0699b459498b72601def19cc4a9ffeabaf9ea91b5017\": container with ID starting with a04066155246b1f7e94c0699b459498b72601def19cc4a9ffeabaf9ea91b5017 not found: ID does not exist" containerID="a04066155246b1f7e94c0699b459498b72601def19cc4a9ffeabaf9ea91b5017" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.014940 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a04066155246b1f7e94c0699b459498b72601def19cc4a9ffeabaf9ea91b5017"} err="failed to get container status \"a04066155246b1f7e94c0699b459498b72601def19cc4a9ffeabaf9ea91b5017\": rpc error: code = NotFound desc = could not find container \"a04066155246b1f7e94c0699b459498b72601def19cc4a9ffeabaf9ea91b5017\": container with ID starting with a04066155246b1f7e94c0699b459498b72601def19cc4a9ffeabaf9ea91b5017 not found: ID does not exist" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.014968 4815 scope.go:117] "RemoveContainer" containerID="52d80bc9c6229451f2e24f36468de30d9ed19786d83228e3b2cbe9b82c63b7ce" Dec 05 09:12:01 crc kubenswrapper[4815]: E1205 09:12:01.015235 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52d80bc9c6229451f2e24f36468de30d9ed19786d83228e3b2cbe9b82c63b7ce\": container with ID starting with 52d80bc9c6229451f2e24f36468de30d9ed19786d83228e3b2cbe9b82c63b7ce not found: ID does not exist" containerID="52d80bc9c6229451f2e24f36468de30d9ed19786d83228e3b2cbe9b82c63b7ce" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.015255 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52d80bc9c6229451f2e24f36468de30d9ed19786d83228e3b2cbe9b82c63b7ce"} err="failed to get container status \"52d80bc9c6229451f2e24f36468de30d9ed19786d83228e3b2cbe9b82c63b7ce\": rpc error: code = NotFound desc = could not find container \"52d80bc9c6229451f2e24f36468de30d9ed19786d83228e3b2cbe9b82c63b7ce\": container with ID starting with 52d80bc9c6229451f2e24f36468de30d9ed19786d83228e3b2cbe9b82c63b7ce not found: ID does not exist" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.015271 4815 scope.go:117] "RemoveContainer" containerID="83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.022977 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" (UID: "e7fa3908-d40f-43d1-9f59-7ad5c8a9877a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.029875 4815 scope.go:117] "RemoveContainer" containerID="83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d" Dec 05 09:12:01 crc kubenswrapper[4815]: E1205 09:12:01.030242 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d\": container with ID starting with 83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d not found: ID does not exist" containerID="83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.030271 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d"} err="failed to get container status \"83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d\": rpc error: code = NotFound desc = could not find container \"83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d\": container with ID starting with 83a6cd83ae4f0e8be8c67befe073a2c25903e98ac5705a6bb21c6946713b9f0d not found: ID does not exist" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.077741 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.083768 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rbpxh"] Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.092078 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rbpxh"] Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.095635 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-kb5lk"] Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.098803 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-kb5lk"] Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.112954 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nntc6"] Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.115756 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nntc6"] Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.134696 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7nc7"] Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.147435 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-k7nc7"] Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.147487 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgk4c"] Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.147517 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-dgk4c"] Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.429452 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" path="/var/lib/kubelet/pods/15a8ad31-81dd-4764-b4da-3c9b3eaa3b25/volumes" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.429935 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="201dec63-11da-4e0d-93ae-025de8a76beb" path="/var/lib/kubelet/pods/201dec63-11da-4e0d-93ae-025de8a76beb/volumes" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.430533 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" path="/var/lib/kubelet/pods/8084d369-d4ee-4dad-bdd3-c8d7ea34199c/volumes" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.431063 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" path="/var/lib/kubelet/pods/aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d/volumes" Dec 05 09:12:01 crc kubenswrapper[4815]: I1205 09:12:01.432193 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" path="/var/lib/kubelet/pods/e7fa3908-d40f-43d1-9f59-7ad5c8a9877a/volumes" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037244 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6k7mn"] Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037424 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037438 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037446 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037451 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037458 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" containerName="extract-utilities" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037466 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" containerName="extract-utilities" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037474 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerName="extract-content" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037481 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerName="extract-content" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037510 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="201dec63-11da-4e0d-93ae-025de8a76beb" containerName="extract-utilities" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037517 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="201dec63-11da-4e0d-93ae-025de8a76beb" containerName="extract-utilities" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037526 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" containerName="marketplace-operator" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037532 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" containerName="marketplace-operator" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037542 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" containerName="extract-content" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037548 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" containerName="extract-content" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037555 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037561 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037568 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037573 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037580 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerName="extract-content" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037586 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerName="extract-content" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037595 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037601 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037612 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" containerName="extract-content" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037619 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" containerName="extract-content" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037626 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="201dec63-11da-4e0d-93ae-025de8a76beb" containerName="extract-content" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037632 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="201dec63-11da-4e0d-93ae-025de8a76beb" containerName="extract-content" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037640 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerName="extract-utilities" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037646 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerName="extract-utilities" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037654 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="201dec63-11da-4e0d-93ae-025de8a76beb" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037660 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="201dec63-11da-4e0d-93ae-025de8a76beb" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037668 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" containerName="extract-utilities" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037674 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" containerName="extract-utilities" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037681 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f986909-b717-447f-be83-09b47b6deb8d" containerName="installer" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037686 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f986909-b717-447f-be83-09b47b6deb8d" containerName="installer" Dec 05 09:12:02 crc kubenswrapper[4815]: E1205 09:12:02.037693 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerName="extract-utilities" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037698 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerName="extract-utilities" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037778 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="15a8ad31-81dd-4764-b4da-3c9b3eaa3b25" containerName="marketplace-operator" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037790 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8084d369-d4ee-4dad-bdd3-c8d7ea34199c" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037799 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa7c866c-ae6f-4bb8-8d9c-4f905fd5d56d" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037807 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d3dc703-3608-44cd-8993-58eaaef64058" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037819 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="201dec63-11da-4e0d-93ae-025de8a76beb" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037827 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7fa3908-d40f-43d1-9f59-7ad5c8a9877a" containerName="registry-server" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037835 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f986909-b717-447f-be83-09b47b6deb8d" containerName="installer" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.037842 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.038170 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.040480 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.041178 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.041653 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.041870 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.053290 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.056173 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6k7mn"] Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.190772 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a4c4b634-bebd-41c4-9462-d33c8a9ff1cd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6k7mn\" (UID: \"a4c4b634-bebd-41c4-9462-d33c8a9ff1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.190870 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9664\" (UniqueName: \"kubernetes.io/projected/a4c4b634-bebd-41c4-9462-d33c8a9ff1cd-kube-api-access-v9664\") pod \"marketplace-operator-79b997595-6k7mn\" (UID: \"a4c4b634-bebd-41c4-9462-d33c8a9ff1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.190895 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4c4b634-bebd-41c4-9462-d33c8a9ff1cd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6k7mn\" (UID: \"a4c4b634-bebd-41c4-9462-d33c8a9ff1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.292098 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9664\" (UniqueName: \"kubernetes.io/projected/a4c4b634-bebd-41c4-9462-d33c8a9ff1cd-kube-api-access-v9664\") pod \"marketplace-operator-79b997595-6k7mn\" (UID: \"a4c4b634-bebd-41c4-9462-d33c8a9ff1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.292142 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4c4b634-bebd-41c4-9462-d33c8a9ff1cd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6k7mn\" (UID: \"a4c4b634-bebd-41c4-9462-d33c8a9ff1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.292164 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a4c4b634-bebd-41c4-9462-d33c8a9ff1cd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6k7mn\" (UID: \"a4c4b634-bebd-41c4-9462-d33c8a9ff1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.294180 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4c4b634-bebd-41c4-9462-d33c8a9ff1cd-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-6k7mn\" (UID: \"a4c4b634-bebd-41c4-9462-d33c8a9ff1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.297156 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/a4c4b634-bebd-41c4-9462-d33c8a9ff1cd-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-6k7mn\" (UID: \"a4c4b634-bebd-41c4-9462-d33c8a9ff1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.311093 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9664\" (UniqueName: \"kubernetes.io/projected/a4c4b634-bebd-41c4-9462-d33c8a9ff1cd-kube-api-access-v9664\") pod \"marketplace-operator-79b997595-6k7mn\" (UID: \"a4c4b634-bebd-41c4-9462-d33c8a9ff1cd\") " pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.356993 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.538935 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-6k7mn"] Dec 05 09:12:02 crc kubenswrapper[4815]: I1205 09:12:02.824641 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" event={"ID":"a4c4b634-bebd-41c4-9462-d33c8a9ff1cd","Type":"ContainerStarted","Data":"262cb06e5d5e1d6e68ffd227ab99be0a79b70364e0df232a53b815c87ac12787"} Dec 05 09:12:03 crc kubenswrapper[4815]: I1205 09:12:03.831225 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" event={"ID":"a4c4b634-bebd-41c4-9462-d33c8a9ff1cd","Type":"ContainerStarted","Data":"834802b33848f100ccdc5939f1f8667d686f54a95b32b80082ead1b6364a3bb1"} Dec 05 09:12:03 crc kubenswrapper[4815]: I1205 09:12:03.831560 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:03 crc kubenswrapper[4815]: I1205 09:12:03.835857 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" Dec 05 09:12:03 crc kubenswrapper[4815]: I1205 09:12:03.852115 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-6k7mn" podStartSLOduration=1.852094871 podStartE2EDuration="1.852094871s" podCreationTimestamp="2025-12-05 09:12:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:12:03.850593358 +0000 UTC m=+322.729200205" watchObservedRunningTime="2025-12-05 09:12:03.852094871 +0000 UTC m=+322.730701708" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.408590 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h66pq"] Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.409140 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" podUID="036169eb-913d-470b-b749-6583deb8f396" containerName="controller-manager" containerID="cri-o://e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e" gracePeriod=30 Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.505662 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8"] Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.505852 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" podUID="cbf444c4-dd3b-4cf7-9771-fea9294124c5" containerName="route-controller-manager" containerID="cri-o://682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04" gracePeriod=30 Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.777830 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.854743 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.893560 4815 generic.go:334] "Generic (PLEG): container finished" podID="036169eb-913d-470b-b749-6583deb8f396" containerID="e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e" exitCode=0 Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.893621 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" event={"ID":"036169eb-913d-470b-b749-6583deb8f396","Type":"ContainerDied","Data":"e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e"} Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.893648 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" event={"ID":"036169eb-913d-470b-b749-6583deb8f396","Type":"ContainerDied","Data":"5de27619c47d116ea3754067b8a9a1cfab543f078cd18721d170bde70d160710"} Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.893663 4815 scope.go:117] "RemoveContainer" containerID="e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.893770 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-h66pq" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.902256 4815 generic.go:334] "Generic (PLEG): container finished" podID="cbf444c4-dd3b-4cf7-9771-fea9294124c5" containerID="682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04" exitCode=0 Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.902306 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" event={"ID":"cbf444c4-dd3b-4cf7-9771-fea9294124c5","Type":"ContainerDied","Data":"682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04"} Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.902337 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" event={"ID":"cbf444c4-dd3b-4cf7-9771-fea9294124c5","Type":"ContainerDied","Data":"eeafab3202013910bc4f98b4e79bb11ef3df20dcbc0e4fe16bd11b3723618bea"} Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.902408 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.926431 4815 scope.go:117] "RemoveContainer" containerID="e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e" Dec 05 09:12:14 crc kubenswrapper[4815]: E1205 09:12:14.927529 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e\": container with ID starting with e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e not found: ID does not exist" containerID="e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.927570 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e"} err="failed to get container status \"e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e\": rpc error: code = NotFound desc = could not find container \"e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e\": container with ID starting with e6d85636e17d936a31d9751c5e465bc07c5e1376f65c93b2ff83fdec2132543e not found: ID does not exist" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.927591 4815 scope.go:117] "RemoveContainer" containerID="682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.941827 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-client-ca\") pod \"036169eb-913d-470b-b749-6583deb8f396\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.941876 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/036169eb-913d-470b-b749-6583deb8f396-serving-cert\") pod \"036169eb-913d-470b-b749-6583deb8f396\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.941916 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6bs4\" (UniqueName: \"kubernetes.io/projected/036169eb-913d-470b-b749-6583deb8f396-kube-api-access-r6bs4\") pod \"036169eb-913d-470b-b749-6583deb8f396\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.941940 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf444c4-dd3b-4cf7-9771-fea9294124c5-serving-cert\") pod \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.941977 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-client-ca\") pod \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.942018 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftkq9\" (UniqueName: \"kubernetes.io/projected/cbf444c4-dd3b-4cf7-9771-fea9294124c5-kube-api-access-ftkq9\") pod \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.942059 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-config\") pod \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\" (UID: \"cbf444c4-dd3b-4cf7-9771-fea9294124c5\") " Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.942084 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-proxy-ca-bundles\") pod \"036169eb-913d-470b-b749-6583deb8f396\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.942106 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-config\") pod \"036169eb-913d-470b-b749-6583deb8f396\" (UID: \"036169eb-913d-470b-b749-6583deb8f396\") " Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.942815 4815 scope.go:117] "RemoveContainer" containerID="682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.942928 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-config" (OuterVolumeSpecName: "config") pod "036169eb-913d-470b-b749-6583deb8f396" (UID: "036169eb-913d-470b-b749-6583deb8f396"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:14 crc kubenswrapper[4815]: E1205 09:12:14.943274 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04\": container with ID starting with 682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04 not found: ID does not exist" containerID="682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.943305 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04"} err="failed to get container status \"682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04\": rpc error: code = NotFound desc = could not find container \"682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04\": container with ID starting with 682fc00c67bf41f02efce8277cd1ea0d58f5255b11ec0c7b5ac75492033efa04 not found: ID does not exist" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.943324 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-client-ca" (OuterVolumeSpecName: "client-ca") pod "cbf444c4-dd3b-4cf7-9771-fea9294124c5" (UID: "cbf444c4-dd3b-4cf7-9771-fea9294124c5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.943702 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-client-ca" (OuterVolumeSpecName: "client-ca") pod "036169eb-913d-470b-b749-6583deb8f396" (UID: "036169eb-913d-470b-b749-6583deb8f396"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.943925 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "036169eb-913d-470b-b749-6583deb8f396" (UID: "036169eb-913d-470b-b749-6583deb8f396"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.943939 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-config" (OuterVolumeSpecName: "config") pod "cbf444c4-dd3b-4cf7-9771-fea9294124c5" (UID: "cbf444c4-dd3b-4cf7-9771-fea9294124c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.948064 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/036169eb-913d-470b-b749-6583deb8f396-kube-api-access-r6bs4" (OuterVolumeSpecName: "kube-api-access-r6bs4") pod "036169eb-913d-470b-b749-6583deb8f396" (UID: "036169eb-913d-470b-b749-6583deb8f396"). InnerVolumeSpecName "kube-api-access-r6bs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.948098 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbf444c4-dd3b-4cf7-9771-fea9294124c5-kube-api-access-ftkq9" (OuterVolumeSpecName: "kube-api-access-ftkq9") pod "cbf444c4-dd3b-4cf7-9771-fea9294124c5" (UID: "cbf444c4-dd3b-4cf7-9771-fea9294124c5"). InnerVolumeSpecName "kube-api-access-ftkq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.948116 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/036169eb-913d-470b-b749-6583deb8f396-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "036169eb-913d-470b-b749-6583deb8f396" (UID: "036169eb-913d-470b-b749-6583deb8f396"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:12:14 crc kubenswrapper[4815]: I1205 09:12:14.948445 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbf444c4-dd3b-4cf7-9771-fea9294124c5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cbf444c4-dd3b-4cf7-9771-fea9294124c5" (UID: "cbf444c4-dd3b-4cf7-9771-fea9294124c5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.044030 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.044078 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/036169eb-913d-470b-b749-6583deb8f396-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.044092 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6bs4\" (UniqueName: \"kubernetes.io/projected/036169eb-913d-470b-b749-6583deb8f396-kube-api-access-r6bs4\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.044106 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbf444c4-dd3b-4cf7-9771-fea9294124c5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.044117 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.044130 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftkq9\" (UniqueName: \"kubernetes.io/projected/cbf444c4-dd3b-4cf7-9771-fea9294124c5-kube-api-access-ftkq9\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.044142 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbf444c4-dd3b-4cf7-9771-fea9294124c5-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.044152 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.044164 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/036169eb-913d-470b-b749-6583deb8f396-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.233697 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h66pq"] Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.236472 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-h66pq"] Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.253258 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8"] Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.256290 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lnmp8"] Dec 05 09:12:15 crc kubenswrapper[4815]: E1205 09:12:15.272896 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod036169eb_913d_470b_b749_6583deb8f396.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbf444c4_dd3b_4cf7_9771_fea9294124c5.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod036169eb_913d_470b_b749_6583deb8f396.slice/crio-5de27619c47d116ea3754067b8a9a1cfab543f078cd18721d170bde70d160710\": RecentStats: unable to find data in memory cache]" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.426738 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="036169eb-913d-470b-b749-6583deb8f396" path="/var/lib/kubelet/pods/036169eb-913d-470b-b749-6583deb8f396/volumes" Dec 05 09:12:15 crc kubenswrapper[4815]: I1205 09:12:15.428202 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbf444c4-dd3b-4cf7-9771-fea9294124c5" path="/var/lib/kubelet/pods/cbf444c4-dd3b-4cf7-9771-fea9294124c5/volumes" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.075391 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh"] Dec 05 09:12:16 crc kubenswrapper[4815]: E1205 09:12:16.075641 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="036169eb-913d-470b-b749-6583deb8f396" containerName="controller-manager" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.075660 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="036169eb-913d-470b-b749-6583deb8f396" containerName="controller-manager" Dec 05 09:12:16 crc kubenswrapper[4815]: E1205 09:12:16.075675 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbf444c4-dd3b-4cf7-9771-fea9294124c5" containerName="route-controller-manager" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.075683 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbf444c4-dd3b-4cf7-9771-fea9294124c5" containerName="route-controller-manager" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.075791 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbf444c4-dd3b-4cf7-9771-fea9294124c5" containerName="route-controller-manager" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.075807 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="036169eb-913d-470b-b749-6583deb8f396" containerName="controller-manager" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.076247 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.079211 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.079531 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.079683 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.079866 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.080000 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.080336 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.083632 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-xwdfc"] Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.084779 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.087925 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.089577 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.089622 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.089874 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.090047 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.090204 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.097606 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh"] Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.112928 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.113323 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-xwdfc"] Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.261639 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpdjp\" (UniqueName: \"kubernetes.io/projected/442eb438-8298-4fa3-b257-62df544770ed-kube-api-access-kpdjp\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.261708 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-client-ca\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.261733 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-proxy-ca-bundles\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.261750 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/442eb438-8298-4fa3-b257-62df544770ed-serving-cert\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.261780 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-config\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.261796 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhl86\" (UniqueName: \"kubernetes.io/projected/fe73a839-053c-4367-94b1-76d71962352f-kube-api-access-vhl86\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.261815 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe73a839-053c-4367-94b1-76d71962352f-serving-cert\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.261840 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-config\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.261857 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-client-ca\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.362646 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpdjp\" (UniqueName: \"kubernetes.io/projected/442eb438-8298-4fa3-b257-62df544770ed-kube-api-access-kpdjp\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.362732 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-client-ca\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.362767 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-proxy-ca-bundles\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.362789 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/442eb438-8298-4fa3-b257-62df544770ed-serving-cert\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.362813 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-config\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.362836 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhl86\" (UniqueName: \"kubernetes.io/projected/fe73a839-053c-4367-94b1-76d71962352f-kube-api-access-vhl86\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.362863 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe73a839-053c-4367-94b1-76d71962352f-serving-cert\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.362893 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-config\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.362916 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-client-ca\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.364509 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-client-ca\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.364564 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-config\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.364746 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-client-ca\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.364860 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-config\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.365034 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-proxy-ca-bundles\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.368660 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/442eb438-8298-4fa3-b257-62df544770ed-serving-cert\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.369657 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe73a839-053c-4367-94b1-76d71962352f-serving-cert\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.391454 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhl86\" (UniqueName: \"kubernetes.io/projected/fe73a839-053c-4367-94b1-76d71962352f-kube-api-access-vhl86\") pod \"controller-manager-58f8979484-xwdfc\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.392603 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpdjp\" (UniqueName: \"kubernetes.io/projected/442eb438-8298-4fa3-b257-62df544770ed-kube-api-access-kpdjp\") pod \"route-controller-manager-596dc78bdd-kjgmh\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.398294 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.418743 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.662499 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-xwdfc"] Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.912701 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" event={"ID":"fe73a839-053c-4367-94b1-76d71962352f","Type":"ContainerStarted","Data":"71b91e232ed220eeb90ac8f6f5cb72808366fd8cfe8e955ae22cd233ef2d065b"} Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.913062 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" event={"ID":"fe73a839-053c-4367-94b1-76d71962352f","Type":"ContainerStarted","Data":"e3e0205bebf9e4d4b7ccca6ed9d1f1885f7b190127bef98658d6a2ac0f0c921c"} Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.913087 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.917397 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.952851 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" podStartSLOduration=2.952833835 podStartE2EDuration="2.952833835s" podCreationTimestamp="2025-12-05 09:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:12:16.949907122 +0000 UTC m=+335.828513949" watchObservedRunningTime="2025-12-05 09:12:16.952833835 +0000 UTC m=+335.831440672" Dec 05 09:12:16 crc kubenswrapper[4815]: I1205 09:12:16.978986 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh"] Dec 05 09:12:17 crc kubenswrapper[4815]: I1205 09:12:17.924034 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" event={"ID":"442eb438-8298-4fa3-b257-62df544770ed","Type":"ContainerStarted","Data":"0fa58d707a0f4ec029e9af91862c204c16aed1628c810e2151d2ccfd35b51bbc"} Dec 05 09:12:17 crc kubenswrapper[4815]: I1205 09:12:17.924368 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:17 crc kubenswrapper[4815]: I1205 09:12:17.924388 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" event={"ID":"442eb438-8298-4fa3-b257-62df544770ed","Type":"ContainerStarted","Data":"08f719c2dfc36a35d673b804a730988b1228e178db6a0a23bb3045fc9a488d9d"} Dec 05 09:12:17 crc kubenswrapper[4815]: I1205 09:12:17.929225 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:17 crc kubenswrapper[4815]: I1205 09:12:17.965119 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" podStartSLOduration=3.965097008 podStartE2EDuration="3.965097008s" podCreationTimestamp="2025-12-05 09:12:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:12:17.948971019 +0000 UTC m=+336.827577876" watchObservedRunningTime="2025-12-05 09:12:17.965097008 +0000 UTC m=+336.843703845" Dec 05 09:12:20 crc kubenswrapper[4815]: I1205 09:12:20.921320 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ssfkz"] Dec 05 09:12:20 crc kubenswrapper[4815]: I1205 09:12:20.923207 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:20 crc kubenswrapper[4815]: I1205 09:12:20.925932 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 05 09:12:20 crc kubenswrapper[4815]: I1205 09:12:20.933362 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ssfkz"] Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.030034 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6be68d8-ee10-4dad-b1d5-5f1d7826917b-catalog-content\") pod \"community-operators-ssfkz\" (UID: \"f6be68d8-ee10-4dad-b1d5-5f1d7826917b\") " pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.030089 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hj24z\" (UniqueName: \"kubernetes.io/projected/f6be68d8-ee10-4dad-b1d5-5f1d7826917b-kube-api-access-hj24z\") pod \"community-operators-ssfkz\" (UID: \"f6be68d8-ee10-4dad-b1d5-5f1d7826917b\") " pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.030109 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6be68d8-ee10-4dad-b1d5-5f1d7826917b-utilities\") pod \"community-operators-ssfkz\" (UID: \"f6be68d8-ee10-4dad-b1d5-5f1d7826917b\") " pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.131636 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6be68d8-ee10-4dad-b1d5-5f1d7826917b-catalog-content\") pod \"community-operators-ssfkz\" (UID: \"f6be68d8-ee10-4dad-b1d5-5f1d7826917b\") " pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.131703 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hj24z\" (UniqueName: \"kubernetes.io/projected/f6be68d8-ee10-4dad-b1d5-5f1d7826917b-kube-api-access-hj24z\") pod \"community-operators-ssfkz\" (UID: \"f6be68d8-ee10-4dad-b1d5-5f1d7826917b\") " pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.131731 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6be68d8-ee10-4dad-b1d5-5f1d7826917b-utilities\") pod \"community-operators-ssfkz\" (UID: \"f6be68d8-ee10-4dad-b1d5-5f1d7826917b\") " pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.132168 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f6be68d8-ee10-4dad-b1d5-5f1d7826917b-catalog-content\") pod \"community-operators-ssfkz\" (UID: \"f6be68d8-ee10-4dad-b1d5-5f1d7826917b\") " pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.132203 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f6be68d8-ee10-4dad-b1d5-5f1d7826917b-utilities\") pod \"community-operators-ssfkz\" (UID: \"f6be68d8-ee10-4dad-b1d5-5f1d7826917b\") " pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.157675 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hj24z\" (UniqueName: \"kubernetes.io/projected/f6be68d8-ee10-4dad-b1d5-5f1d7826917b-kube-api-access-hj24z\") pod \"community-operators-ssfkz\" (UID: \"f6be68d8-ee10-4dad-b1d5-5f1d7826917b\") " pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.244808 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.676552 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ssfkz"] Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.943375 4815 generic.go:334] "Generic (PLEG): container finished" podID="f6be68d8-ee10-4dad-b1d5-5f1d7826917b" containerID="4188f7aa90e2d5b817c8b9ded2267240cf008f46f89738b996b8490b9973ecdf" exitCode=0 Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.943416 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssfkz" event={"ID":"f6be68d8-ee10-4dad-b1d5-5f1d7826917b","Type":"ContainerDied","Data":"4188f7aa90e2d5b817c8b9ded2267240cf008f46f89738b996b8490b9973ecdf"} Dec 05 09:12:21 crc kubenswrapper[4815]: I1205 09:12:21.943439 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssfkz" event={"ID":"f6be68d8-ee10-4dad-b1d5-5f1d7826917b","Type":"ContainerStarted","Data":"f54537e57b2e701d3a3b85afba6862c9c353d4c7faec6b7bb38e1a1a34edf5e6"} Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.094984 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-xwdfc"] Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.095250 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" podUID="fe73a839-053c-4367-94b1-76d71962352f" containerName="controller-manager" containerID="cri-o://71b91e232ed220eeb90ac8f6f5cb72808366fd8cfe8e955ae22cd233ef2d065b" gracePeriod=30 Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.113159 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh"] Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.113424 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" podUID="442eb438-8298-4fa3-b257-62df544770ed" containerName="route-controller-manager" containerID="cri-o://0fa58d707a0f4ec029e9af91862c204c16aed1628c810e2151d2ccfd35b51bbc" gracePeriod=30 Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.729684 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-4lpmz"] Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.730891 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:22 crc kubenswrapper[4815]: W1205 09:12:22.735762 4815 reflector.go:561] object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb": failed to list *v1.Secret: secrets "redhat-marketplace-dockercfg-x2ctb" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 05 09:12:22 crc kubenswrapper[4815]: E1205 09:12:22.735806 4815 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"redhat-marketplace-dockercfg-x2ctb\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"redhat-marketplace-dockercfg-x2ctb\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.822837 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dxvr\" (UniqueName: \"kubernetes.io/projected/d66535f7-8f1a-48cd-af4d-5c6a41535fe5-kube-api-access-9dxvr\") pod \"redhat-marketplace-4lpmz\" (UID: \"d66535f7-8f1a-48cd-af4d-5c6a41535fe5\") " pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.823182 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66535f7-8f1a-48cd-af4d-5c6a41535fe5-catalog-content\") pod \"redhat-marketplace-4lpmz\" (UID: \"d66535f7-8f1a-48cd-af4d-5c6a41535fe5\") " pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.823284 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66535f7-8f1a-48cd-af4d-5c6a41535fe5-utilities\") pod \"redhat-marketplace-4lpmz\" (UID: \"d66535f7-8f1a-48cd-af4d-5c6a41535fe5\") " pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.828669 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lpmz"] Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.924225 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66535f7-8f1a-48cd-af4d-5c6a41535fe5-utilities\") pod \"redhat-marketplace-4lpmz\" (UID: \"d66535f7-8f1a-48cd-af4d-5c6a41535fe5\") " pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.924307 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dxvr\" (UniqueName: \"kubernetes.io/projected/d66535f7-8f1a-48cd-af4d-5c6a41535fe5-kube-api-access-9dxvr\") pod \"redhat-marketplace-4lpmz\" (UID: \"d66535f7-8f1a-48cd-af4d-5c6a41535fe5\") " pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.924358 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66535f7-8f1a-48cd-af4d-5c6a41535fe5-catalog-content\") pod \"redhat-marketplace-4lpmz\" (UID: \"d66535f7-8f1a-48cd-af4d-5c6a41535fe5\") " pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.924858 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d66535f7-8f1a-48cd-af4d-5c6a41535fe5-catalog-content\") pod \"redhat-marketplace-4lpmz\" (UID: \"d66535f7-8f1a-48cd-af4d-5c6a41535fe5\") " pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.925138 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d66535f7-8f1a-48cd-af4d-5c6a41535fe5-utilities\") pod \"redhat-marketplace-4lpmz\" (UID: \"d66535f7-8f1a-48cd-af4d-5c6a41535fe5\") " pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.951608 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dxvr\" (UniqueName: \"kubernetes.io/projected/d66535f7-8f1a-48cd-af4d-5c6a41535fe5-kube-api-access-9dxvr\") pod \"redhat-marketplace-4lpmz\" (UID: \"d66535f7-8f1a-48cd-af4d-5c6a41535fe5\") " pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.962819 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssfkz" event={"ID":"f6be68d8-ee10-4dad-b1d5-5f1d7826917b","Type":"ContainerStarted","Data":"b9bd1eea9040ee273c8655000334972d344d3a41a7d0110662e7ca184e12633c"} Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.969102 4815 generic.go:334] "Generic (PLEG): container finished" podID="442eb438-8298-4fa3-b257-62df544770ed" containerID="0fa58d707a0f4ec029e9af91862c204c16aed1628c810e2151d2ccfd35b51bbc" exitCode=0 Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.969175 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" event={"ID":"442eb438-8298-4fa3-b257-62df544770ed","Type":"ContainerDied","Data":"0fa58d707a0f4ec029e9af91862c204c16aed1628c810e2151d2ccfd35b51bbc"} Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.973838 4815 generic.go:334] "Generic (PLEG): container finished" podID="fe73a839-053c-4367-94b1-76d71962352f" containerID="71b91e232ed220eeb90ac8f6f5cb72808366fd8cfe8e955ae22cd233ef2d065b" exitCode=0 Dec 05 09:12:22 crc kubenswrapper[4815]: I1205 09:12:22.973894 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" event={"ID":"fe73a839-053c-4367-94b1-76d71962352f","Type":"ContainerDied","Data":"71b91e232ed220eeb90ac8f6f5cb72808366fd8cfe8e955ae22cd233ef2d065b"} Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.421586 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.427731 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.430673 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-proxy-ca-bundles\") pod \"fe73a839-053c-4367-94b1-76d71962352f\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.430708 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kpdjp\" (UniqueName: \"kubernetes.io/projected/442eb438-8298-4fa3-b257-62df544770ed-kube-api-access-kpdjp\") pod \"442eb438-8298-4fa3-b257-62df544770ed\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.430732 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-config\") pod \"fe73a839-053c-4367-94b1-76d71962352f\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.430759 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhl86\" (UniqueName: \"kubernetes.io/projected/fe73a839-053c-4367-94b1-76d71962352f-kube-api-access-vhl86\") pod \"fe73a839-053c-4367-94b1-76d71962352f\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.430778 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe73a839-053c-4367-94b1-76d71962352f-serving-cert\") pod \"fe73a839-053c-4367-94b1-76d71962352f\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.430815 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-config\") pod \"442eb438-8298-4fa3-b257-62df544770ed\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.430830 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-client-ca\") pod \"fe73a839-053c-4367-94b1-76d71962352f\" (UID: \"fe73a839-053c-4367-94b1-76d71962352f\") " Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.430857 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-client-ca\") pod \"442eb438-8298-4fa3-b257-62df544770ed\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.430874 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/442eb438-8298-4fa3-b257-62df544770ed-serving-cert\") pod \"442eb438-8298-4fa3-b257-62df544770ed\" (UID: \"442eb438-8298-4fa3-b257-62df544770ed\") " Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.432280 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-config" (OuterVolumeSpecName: "config") pod "fe73a839-053c-4367-94b1-76d71962352f" (UID: "fe73a839-053c-4367-94b1-76d71962352f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.432313 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-client-ca" (OuterVolumeSpecName: "client-ca") pod "442eb438-8298-4fa3-b257-62df544770ed" (UID: "442eb438-8298-4fa3-b257-62df544770ed"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.432408 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-client-ca" (OuterVolumeSpecName: "client-ca") pod "fe73a839-053c-4367-94b1-76d71962352f" (UID: "fe73a839-053c-4367-94b1-76d71962352f"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.432418 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-config" (OuterVolumeSpecName: "config") pod "442eb438-8298-4fa3-b257-62df544770ed" (UID: "442eb438-8298-4fa3-b257-62df544770ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.432425 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "fe73a839-053c-4367-94b1-76d71962352f" (UID: "fe73a839-053c-4367-94b1-76d71962352f"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.434880 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/442eb438-8298-4fa3-b257-62df544770ed-kube-api-access-kpdjp" (OuterVolumeSpecName: "kube-api-access-kpdjp") pod "442eb438-8298-4fa3-b257-62df544770ed" (UID: "442eb438-8298-4fa3-b257-62df544770ed"). InnerVolumeSpecName "kube-api-access-kpdjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.435040 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe73a839-053c-4367-94b1-76d71962352f-kube-api-access-vhl86" (OuterVolumeSpecName: "kube-api-access-vhl86") pod "fe73a839-053c-4367-94b1-76d71962352f" (UID: "fe73a839-053c-4367-94b1-76d71962352f"). InnerVolumeSpecName "kube-api-access-vhl86". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.435048 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/442eb438-8298-4fa3-b257-62df544770ed-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "442eb438-8298-4fa3-b257-62df544770ed" (UID: "442eb438-8298-4fa3-b257-62df544770ed"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.439478 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fe73a839-053c-4367-94b1-76d71962352f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fe73a839-053c-4367-94b1-76d71962352f" (UID: "fe73a839-053c-4367-94b1-76d71962352f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.515757 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-t2dbx"] Dec 05 09:12:23 crc kubenswrapper[4815]: E1205 09:12:23.515956 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe73a839-053c-4367-94b1-76d71962352f" containerName="controller-manager" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.515967 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe73a839-053c-4367-94b1-76d71962352f" containerName="controller-manager" Dec 05 09:12:23 crc kubenswrapper[4815]: E1205 09:12:23.515989 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="442eb438-8298-4fa3-b257-62df544770ed" containerName="route-controller-manager" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.515995 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="442eb438-8298-4fa3-b257-62df544770ed" containerName="route-controller-manager" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.516082 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="442eb438-8298-4fa3-b257-62df544770ed" containerName="route-controller-manager" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.516090 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe73a839-053c-4367-94b1-76d71962352f" containerName="controller-manager" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.523860 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.526795 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.529013 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t2dbx"] Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.534846 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.534892 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.534905 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/442eb438-8298-4fa3-b257-62df544770ed-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.534915 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/442eb438-8298-4fa3-b257-62df544770ed-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.534926 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.534940 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kpdjp\" (UniqueName: \"kubernetes.io/projected/442eb438-8298-4fa3-b257-62df544770ed-kube-api-access-kpdjp\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.534951 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fe73a839-053c-4367-94b1-76d71962352f-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.534963 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhl86\" (UniqueName: \"kubernetes.io/projected/fe73a839-053c-4367-94b1-76d71962352f-kube-api-access-vhl86\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.534973 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fe73a839-053c-4367-94b1-76d71962352f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.635612 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kjbt\" (UniqueName: \"kubernetes.io/projected/18bd7707-24bd-4562-a633-c24cd47e50d0-kube-api-access-5kjbt\") pod \"redhat-operators-t2dbx\" (UID: \"18bd7707-24bd-4562-a633-c24cd47e50d0\") " pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.635666 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18bd7707-24bd-4562-a633-c24cd47e50d0-utilities\") pod \"redhat-operators-t2dbx\" (UID: \"18bd7707-24bd-4562-a633-c24cd47e50d0\") " pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.635745 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18bd7707-24bd-4562-a633-c24cd47e50d0-catalog-content\") pod \"redhat-operators-t2dbx\" (UID: \"18bd7707-24bd-4562-a633-c24cd47e50d0\") " pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.737001 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18bd7707-24bd-4562-a633-c24cd47e50d0-utilities\") pod \"redhat-operators-t2dbx\" (UID: \"18bd7707-24bd-4562-a633-c24cd47e50d0\") " pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.737078 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18bd7707-24bd-4562-a633-c24cd47e50d0-catalog-content\") pod \"redhat-operators-t2dbx\" (UID: \"18bd7707-24bd-4562-a633-c24cd47e50d0\") " pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.737155 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kjbt\" (UniqueName: \"kubernetes.io/projected/18bd7707-24bd-4562-a633-c24cd47e50d0-kube-api-access-5kjbt\") pod \"redhat-operators-t2dbx\" (UID: \"18bd7707-24bd-4562-a633-c24cd47e50d0\") " pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.737825 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18bd7707-24bd-4562-a633-c24cd47e50d0-utilities\") pod \"redhat-operators-t2dbx\" (UID: \"18bd7707-24bd-4562-a633-c24cd47e50d0\") " pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.737835 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18bd7707-24bd-4562-a633-c24cd47e50d0-catalog-content\") pod \"redhat-operators-t2dbx\" (UID: \"18bd7707-24bd-4562-a633-c24cd47e50d0\") " pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.754308 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kjbt\" (UniqueName: \"kubernetes.io/projected/18bd7707-24bd-4562-a633-c24cd47e50d0-kube-api-access-5kjbt\") pod \"redhat-operators-t2dbx\" (UID: \"18bd7707-24bd-4562-a633-c24cd47e50d0\") " pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.851876 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.897670 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 05 09:12:23 crc kubenswrapper[4815]: I1205 09:12:23.901577 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.022833 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.022929 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-58f8979484-xwdfc" event={"ID":"fe73a839-053c-4367-94b1-76d71962352f","Type":"ContainerDied","Data":"e3e0205bebf9e4d4b7ccca6ed9d1f1885f7b190127bef98658d6a2ac0f0c921c"} Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.022964 4815 scope.go:117] "RemoveContainer" containerID="71b91e232ed220eeb90ac8f6f5cb72808366fd8cfe8e955ae22cd233ef2d065b" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.035047 4815 generic.go:334] "Generic (PLEG): container finished" podID="f6be68d8-ee10-4dad-b1d5-5f1d7826917b" containerID="b9bd1eea9040ee273c8655000334972d344d3a41a7d0110662e7ca184e12633c" exitCode=0 Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.035419 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssfkz" event={"ID":"f6be68d8-ee10-4dad-b1d5-5f1d7826917b","Type":"ContainerDied","Data":"b9bd1eea9040ee273c8655000334972d344d3a41a7d0110662e7ca184e12633c"} Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.043974 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" event={"ID":"442eb438-8298-4fa3-b257-62df544770ed","Type":"ContainerDied","Data":"08f719c2dfc36a35d673b804a730988b1228e178db6a0a23bb3045fc9a488d9d"} Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.044025 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.085667 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-xwdfc"] Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.086849 4815 scope.go:117] "RemoveContainer" containerID="0fa58d707a0f4ec029e9af91862c204c16aed1628c810e2151d2ccfd35b51bbc" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.104881 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-58f8979484-xwdfc"] Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.113651 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5964cbcb45-6ckhp"] Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.114746 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.120840 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.121055 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.121108 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.121303 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.123081 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.123244 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.125465 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst"] Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.127097 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.145484 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5964cbcb45-6ckhp"] Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.145534 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst"] Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.145612 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.149169 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh"] Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.157955 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.158160 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.158285 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.158411 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.161146 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-596dc78bdd-kjgmh"] Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.161713 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.171170 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.189242 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-4lpmz"] Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.244612 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-config\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.244688 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n59m\" (UniqueName: \"kubernetes.io/projected/cbccfaaf-566f-4e91-84d0-a18b86e848cc-kube-api-access-9n59m\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.244717 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-client-ca\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.244746 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbccfaaf-566f-4e91-84d0-a18b86e848cc-serving-cert\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.244794 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-proxy-ca-bundles\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.244824 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7fdrg\" (UniqueName: \"kubernetes.io/projected/4b1e66e9-1c8c-4392-9a04-185c83064989-kube-api-access-7fdrg\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.244852 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b1e66e9-1c8c-4392-9a04-185c83064989-serving-cert\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.244900 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-config\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.244920 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-client-ca\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.256447 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-t2dbx"] Dec 05 09:12:24 crc kubenswrapper[4815]: W1205 09:12:24.263252 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod18bd7707_24bd_4562_a633_c24cd47e50d0.slice/crio-047ff671b89d6721f8a1d5b7fc63fb2692d0fad4852c7950e7b5cc4b5ac8f271 WatchSource:0}: Error finding container 047ff671b89d6721f8a1d5b7fc63fb2692d0fad4852c7950e7b5cc4b5ac8f271: Status 404 returned error can't find the container with id 047ff671b89d6721f8a1d5b7fc63fb2692d0fad4852c7950e7b5cc4b5ac8f271 Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.345730 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n59m\" (UniqueName: \"kubernetes.io/projected/cbccfaaf-566f-4e91-84d0-a18b86e848cc-kube-api-access-9n59m\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.345813 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-client-ca\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.345852 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbccfaaf-566f-4e91-84d0-a18b86e848cc-serving-cert\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.345886 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-proxy-ca-bundles\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.345908 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7fdrg\" (UniqueName: \"kubernetes.io/projected/4b1e66e9-1c8c-4392-9a04-185c83064989-kube-api-access-7fdrg\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.345929 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b1e66e9-1c8c-4392-9a04-185c83064989-serving-cert\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.345944 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-config\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.345963 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-client-ca\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.345998 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-config\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.348475 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-proxy-ca-bundles\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.351011 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-client-ca\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.351075 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-client-ca\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.351328 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-config\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.352149 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-config\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.356796 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbccfaaf-566f-4e91-84d0-a18b86e848cc-serving-cert\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.357647 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b1e66e9-1c8c-4392-9a04-185c83064989-serving-cert\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.375411 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7fdrg\" (UniqueName: \"kubernetes.io/projected/4b1e66e9-1c8c-4392-9a04-185c83064989-kube-api-access-7fdrg\") pod \"route-controller-manager-5598468cdf-79hst\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.376160 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n59m\" (UniqueName: \"kubernetes.io/projected/cbccfaaf-566f-4e91-84d0-a18b86e848cc-kube-api-access-9n59m\") pod \"controller-manager-5964cbcb45-6ckhp\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.445651 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.480850 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.839156 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst"] Dec 05 09:12:24 crc kubenswrapper[4815]: I1205 09:12:24.864300 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5964cbcb45-6ckhp"] Dec 05 09:12:24 crc kubenswrapper[4815]: W1205 09:12:24.871535 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcbccfaaf_566f_4e91_84d0_a18b86e848cc.slice/crio-5ca6eb659fab3317012a199b3096494a891e15e1c05a47975acd3732a40fcb3b WatchSource:0}: Error finding container 5ca6eb659fab3317012a199b3096494a891e15e1c05a47975acd3732a40fcb3b: Status 404 returned error can't find the container with id 5ca6eb659fab3317012a199b3096494a891e15e1c05a47975acd3732a40fcb3b Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.051699 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" event={"ID":"cbccfaaf-566f-4e91-84d0-a18b86e848cc","Type":"ContainerStarted","Data":"12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005"} Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.051748 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" event={"ID":"cbccfaaf-566f-4e91-84d0-a18b86e848cc","Type":"ContainerStarted","Data":"5ca6eb659fab3317012a199b3096494a891e15e1c05a47975acd3732a40fcb3b"} Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.052904 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.054211 4815 patch_prober.go:28] interesting pod/controller-manager-5964cbcb45-6ckhp container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" start-of-body= Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.054242 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" podUID="cbccfaaf-566f-4e91-84d0-a18b86e848cc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.056998 4815 generic.go:334] "Generic (PLEG): container finished" podID="18bd7707-24bd-4562-a633-c24cd47e50d0" containerID="1442f514d0e8c6b78aa5627bfe3d014d729aa4377080c30d47c2eecc04009668" exitCode=0 Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.057654 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2dbx" event={"ID":"18bd7707-24bd-4562-a633-c24cd47e50d0","Type":"ContainerDied","Data":"1442f514d0e8c6b78aa5627bfe3d014d729aa4377080c30d47c2eecc04009668"} Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.057744 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2dbx" event={"ID":"18bd7707-24bd-4562-a633-c24cd47e50d0","Type":"ContainerStarted","Data":"047ff671b89d6721f8a1d5b7fc63fb2692d0fad4852c7950e7b5cc4b5ac8f271"} Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.059747 4815 generic.go:334] "Generic (PLEG): container finished" podID="d66535f7-8f1a-48cd-af4d-5c6a41535fe5" containerID="80bd2e18b5c4498968a62865cd01d678b284b0c4bbac767205a8244d1b7ad705" exitCode=0 Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.060012 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lpmz" event={"ID":"d66535f7-8f1a-48cd-af4d-5c6a41535fe5","Type":"ContainerDied","Data":"80bd2e18b5c4498968a62865cd01d678b284b0c4bbac767205a8244d1b7ad705"} Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.060043 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lpmz" event={"ID":"d66535f7-8f1a-48cd-af4d-5c6a41535fe5","Type":"ContainerStarted","Data":"c6c14a02412307d93856bca4f48807f3bbb0f2c11fdf14a4cc9803656b41c351"} Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.067100 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ssfkz" event={"ID":"f6be68d8-ee10-4dad-b1d5-5f1d7826917b","Type":"ContainerStarted","Data":"483b11512639989aa624b006ae78271936eba0794836bc5d072abb52efa1f3e6"} Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.077839 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" podStartSLOduration=3.077820513 podStartE2EDuration="3.077820513s" podCreationTimestamp="2025-12-05 09:12:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:12:25.077480693 +0000 UTC m=+343.956087530" watchObservedRunningTime="2025-12-05 09:12:25.077820513 +0000 UTC m=+343.956427350" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.077954 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" event={"ID":"4b1e66e9-1c8c-4392-9a04-185c83064989","Type":"ContainerStarted","Data":"cdafd09012ba11709b36dd28afabbe54f9d0beef3231988f21047706af291b06"} Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.077989 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" event={"ID":"4b1e66e9-1c8c-4392-9a04-185c83064989","Type":"ContainerStarted","Data":"a60d05731ffdbdb3bc5c90a17ad7b1184116a7dfe4d30d516df6116cb4302ce7"} Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.078756 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.136343 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ssfkz" podStartSLOduration=2.56880523 podStartE2EDuration="5.136321164s" podCreationTimestamp="2025-12-05 09:12:20 +0000 UTC" firstStartedPulling="2025-12-05 09:12:21.944789521 +0000 UTC m=+340.823396358" lastFinishedPulling="2025-12-05 09:12:24.512305455 +0000 UTC m=+343.390912292" observedRunningTime="2025-12-05 09:12:25.130633183 +0000 UTC m=+344.009240020" watchObservedRunningTime="2025-12-05 09:12:25.136321164 +0000 UTC m=+344.014928001" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.151452 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" podStartSLOduration=3.151428683 podStartE2EDuration="3.151428683s" podCreationTimestamp="2025-12-05 09:12:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:12:25.146676718 +0000 UTC m=+344.025283555" watchObservedRunningTime="2025-12-05 09:12:25.151428683 +0000 UTC m=+344.030035520" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.325002 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-l4v4h"] Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.326268 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.328572 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.377712 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l4v4h"] Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.442299 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="442eb438-8298-4fa3-b257-62df544770ed" path="/var/lib/kubelet/pods/442eb438-8298-4fa3-b257-62df544770ed/volumes" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.443231 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe73a839-053c-4367-94b1-76d71962352f" path="/var/lib/kubelet/pods/fe73a839-053c-4367-94b1-76d71962352f/volumes" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.465444 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50a2f4af-33e9-4508-853b-6142ca44e9ef-catalog-content\") pod \"certified-operators-l4v4h\" (UID: \"50a2f4af-33e9-4508-853b-6142ca44e9ef\") " pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.465968 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bvlq\" (UniqueName: \"kubernetes.io/projected/50a2f4af-33e9-4508-853b-6142ca44e9ef-kube-api-access-9bvlq\") pod \"certified-operators-l4v4h\" (UID: \"50a2f4af-33e9-4508-853b-6142ca44e9ef\") " pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.466108 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50a2f4af-33e9-4508-853b-6142ca44e9ef-utilities\") pod \"certified-operators-l4v4h\" (UID: \"50a2f4af-33e9-4508-853b-6142ca44e9ef\") " pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.567589 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50a2f4af-33e9-4508-853b-6142ca44e9ef-catalog-content\") pod \"certified-operators-l4v4h\" (UID: \"50a2f4af-33e9-4508-853b-6142ca44e9ef\") " pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.567976 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bvlq\" (UniqueName: \"kubernetes.io/projected/50a2f4af-33e9-4508-853b-6142ca44e9ef-kube-api-access-9bvlq\") pod \"certified-operators-l4v4h\" (UID: \"50a2f4af-33e9-4508-853b-6142ca44e9ef\") " pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.568030 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50a2f4af-33e9-4508-853b-6142ca44e9ef-utilities\") pod \"certified-operators-l4v4h\" (UID: \"50a2f4af-33e9-4508-853b-6142ca44e9ef\") " pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.568614 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/50a2f4af-33e9-4508-853b-6142ca44e9ef-utilities\") pod \"certified-operators-l4v4h\" (UID: \"50a2f4af-33e9-4508-853b-6142ca44e9ef\") " pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.568912 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/50a2f4af-33e9-4508-853b-6142ca44e9ef-catalog-content\") pod \"certified-operators-l4v4h\" (UID: \"50a2f4af-33e9-4508-853b-6142ca44e9ef\") " pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.589397 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bvlq\" (UniqueName: \"kubernetes.io/projected/50a2f4af-33e9-4508-853b-6142ca44e9ef-kube-api-access-9bvlq\") pod \"certified-operators-l4v4h\" (UID: \"50a2f4af-33e9-4508-853b-6142ca44e9ef\") " pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:25 crc kubenswrapper[4815]: I1205 09:12:25.736087 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:26 crc kubenswrapper[4815]: I1205 09:12:26.049315 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:26 crc kubenswrapper[4815]: I1205 09:12:26.124320 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lpmz" event={"ID":"d66535f7-8f1a-48cd-af4d-5c6a41535fe5","Type":"ContainerStarted","Data":"1638fac76ae0e0abfb3f71acc3ac129894480d297b57921ae45a9da1986c605c"} Dec 05 09:12:26 crc kubenswrapper[4815]: I1205 09:12:26.136049 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2dbx" event={"ID":"18bd7707-24bd-4562-a633-c24cd47e50d0","Type":"ContainerStarted","Data":"266cc9e3e6ef42945e43cc706d05b840e27948348b6a7e100909da37197e993a"} Dec 05 09:12:26 crc kubenswrapper[4815]: I1205 09:12:26.146819 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:26 crc kubenswrapper[4815]: I1205 09:12:26.276845 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-l4v4h"] Dec 05 09:12:27 crc kubenswrapper[4815]: I1205 09:12:27.141816 4815 generic.go:334] "Generic (PLEG): container finished" podID="d66535f7-8f1a-48cd-af4d-5c6a41535fe5" containerID="1638fac76ae0e0abfb3f71acc3ac129894480d297b57921ae45a9da1986c605c" exitCode=0 Dec 05 09:12:27 crc kubenswrapper[4815]: I1205 09:12:27.143189 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lpmz" event={"ID":"d66535f7-8f1a-48cd-af4d-5c6a41535fe5","Type":"ContainerDied","Data":"1638fac76ae0e0abfb3f71acc3ac129894480d297b57921ae45a9da1986c605c"} Dec 05 09:12:27 crc kubenswrapper[4815]: I1205 09:12:27.144610 4815 generic.go:334] "Generic (PLEG): container finished" podID="50a2f4af-33e9-4508-853b-6142ca44e9ef" containerID="e9ee8f21d6a48c758b4b60a5a2cbdbab3a875e59c170de9f8d08aaa425c3126b" exitCode=0 Dec 05 09:12:27 crc kubenswrapper[4815]: I1205 09:12:27.145005 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l4v4h" event={"ID":"50a2f4af-33e9-4508-853b-6142ca44e9ef","Type":"ContainerDied","Data":"e9ee8f21d6a48c758b4b60a5a2cbdbab3a875e59c170de9f8d08aaa425c3126b"} Dec 05 09:12:27 crc kubenswrapper[4815]: I1205 09:12:27.145312 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l4v4h" event={"ID":"50a2f4af-33e9-4508-853b-6142ca44e9ef","Type":"ContainerStarted","Data":"52f39c5592d5ff93281d36b93030e530feaf2ed93d5b0899eb193e9db3a4b279"} Dec 05 09:12:28 crc kubenswrapper[4815]: I1205 09:12:28.153007 4815 generic.go:334] "Generic (PLEG): container finished" podID="18bd7707-24bd-4562-a633-c24cd47e50d0" containerID="266cc9e3e6ef42945e43cc706d05b840e27948348b6a7e100909da37197e993a" exitCode=0 Dec 05 09:12:28 crc kubenswrapper[4815]: I1205 09:12:28.153074 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2dbx" event={"ID":"18bd7707-24bd-4562-a633-c24cd47e50d0","Type":"ContainerDied","Data":"266cc9e3e6ef42945e43cc706d05b840e27948348b6a7e100909da37197e993a"} Dec 05 09:12:29 crc kubenswrapper[4815]: I1205 09:12:29.160610 4815 generic.go:334] "Generic (PLEG): container finished" podID="50a2f4af-33e9-4508-853b-6142ca44e9ef" containerID="b10db287dd6bf06e0d8b263d387e475c9e3375eedafb4f370c07b5d4a1a629f7" exitCode=0 Dec 05 09:12:29 crc kubenswrapper[4815]: I1205 09:12:29.160687 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l4v4h" event={"ID":"50a2f4af-33e9-4508-853b-6142ca44e9ef","Type":"ContainerDied","Data":"b10db287dd6bf06e0d8b263d387e475c9e3375eedafb4f370c07b5d4a1a629f7"} Dec 05 09:12:29 crc kubenswrapper[4815]: I1205 09:12:29.162805 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-t2dbx" event={"ID":"18bd7707-24bd-4562-a633-c24cd47e50d0","Type":"ContainerStarted","Data":"f9c9dd7926d476fe937246a35f6eda35eca91ff3bda3af53abfb397e32464870"} Dec 05 09:12:29 crc kubenswrapper[4815]: I1205 09:12:29.164853 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-4lpmz" event={"ID":"d66535f7-8f1a-48cd-af4d-5c6a41535fe5","Type":"ContainerStarted","Data":"8209ee54bac351645acb0a7007d1385502e4748d1d72659254f80fd248c0390f"} Dec 05 09:12:29 crc kubenswrapper[4815]: I1205 09:12:29.205370 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-4lpmz" podStartSLOduration=3.894980296 podStartE2EDuration="7.205348713s" podCreationTimestamp="2025-12-05 09:12:22 +0000 UTC" firstStartedPulling="2025-12-05 09:12:25.065213535 +0000 UTC m=+343.943820372" lastFinishedPulling="2025-12-05 09:12:28.375581952 +0000 UTC m=+347.254188789" observedRunningTime="2025-12-05 09:12:29.201675539 +0000 UTC m=+348.080282396" watchObservedRunningTime="2025-12-05 09:12:29.205348713 +0000 UTC m=+348.083955550" Dec 05 09:12:29 crc kubenswrapper[4815]: I1205 09:12:29.221556 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-t2dbx" podStartSLOduration=2.713007239 podStartE2EDuration="6.221529653s" podCreationTimestamp="2025-12-05 09:12:23 +0000 UTC" firstStartedPulling="2025-12-05 09:12:25.058836904 +0000 UTC m=+343.937443761" lastFinishedPulling="2025-12-05 09:12:28.567359328 +0000 UTC m=+347.445966175" observedRunningTime="2025-12-05 09:12:29.218375033 +0000 UTC m=+348.096981890" watchObservedRunningTime="2025-12-05 09:12:29.221529653 +0000 UTC m=+348.100136500" Dec 05 09:12:30 crc kubenswrapper[4815]: I1205 09:12:30.175378 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-l4v4h" event={"ID":"50a2f4af-33e9-4508-853b-6142ca44e9ef","Type":"ContainerStarted","Data":"b103643b044f7f77daec85678e3c1ce84ba0b23e538719d4283baf804341edb3"} Dec 05 09:12:30 crc kubenswrapper[4815]: I1205 09:12:30.207226 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-l4v4h" podStartSLOduration=2.698059534 podStartE2EDuration="5.207205111s" podCreationTimestamp="2025-12-05 09:12:25 +0000 UTC" firstStartedPulling="2025-12-05 09:12:27.147427069 +0000 UTC m=+346.026033916" lastFinishedPulling="2025-12-05 09:12:29.656572666 +0000 UTC m=+348.535179493" observedRunningTime="2025-12-05 09:12:30.203200838 +0000 UTC m=+349.081807675" watchObservedRunningTime="2025-12-05 09:12:30.207205111 +0000 UTC m=+349.085811948" Dec 05 09:12:31 crc kubenswrapper[4815]: I1205 09:12:31.245255 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:31 crc kubenswrapper[4815]: I1205 09:12:31.245316 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:31 crc kubenswrapper[4815]: I1205 09:12:31.289920 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:32 crc kubenswrapper[4815]: I1205 09:12:32.229669 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ssfkz" Dec 05 09:12:33 crc kubenswrapper[4815]: I1205 09:12:33.852791 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:33 crc kubenswrapper[4815]: I1205 09:12:33.852851 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:33 crc kubenswrapper[4815]: I1205 09:12:33.902266 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:33 crc kubenswrapper[4815]: I1205 09:12:33.902318 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:33 crc kubenswrapper[4815]: I1205 09:12:33.945961 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:34 crc kubenswrapper[4815]: I1205 09:12:34.229116 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-4lpmz" Dec 05 09:12:34 crc kubenswrapper[4815]: I1205 09:12:34.404422 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5964cbcb45-6ckhp"] Dec 05 09:12:34 crc kubenswrapper[4815]: I1205 09:12:34.404950 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" podUID="cbccfaaf-566f-4e91-84d0-a18b86e848cc" containerName="controller-manager" containerID="cri-o://12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005" gracePeriod=30 Dec 05 09:12:34 crc kubenswrapper[4815]: I1205 09:12:34.447521 4815 patch_prober.go:28] interesting pod/controller-manager-5964cbcb45-6ckhp container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" start-of-body= Dec 05 09:12:34 crc kubenswrapper[4815]: I1205 09:12:34.447621 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" podUID="cbccfaaf-566f-4e91-84d0-a18b86e848cc" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.64:8443/healthz\": dial tcp 10.217.0.64:8443: connect: connection refused" Dec 05 09:12:34 crc kubenswrapper[4815]: I1205 09:12:34.891542 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-t2dbx" podUID="18bd7707-24bd-4562-a633-c24cd47e50d0" containerName="registry-server" probeResult="failure" output=< Dec 05 09:12:34 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 09:12:34 crc kubenswrapper[4815]: > Dec 05 09:12:35 crc kubenswrapper[4815]: I1205 09:12:35.736245 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:35 crc kubenswrapper[4815]: I1205 09:12:35.736296 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:35 crc kubenswrapper[4815]: I1205 09:12:35.781324 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:36 crc kubenswrapper[4815]: I1205 09:12:36.239090 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-l4v4h" Dec 05 09:12:38 crc kubenswrapper[4815]: I1205 09:12:38.984798 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.013575 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-68bbc847b9-998jg"] Dec 05 09:12:39 crc kubenswrapper[4815]: E1205 09:12:39.013830 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbccfaaf-566f-4e91-84d0-a18b86e848cc" containerName="controller-manager" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.013849 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbccfaaf-566f-4e91-84d0-a18b86e848cc" containerName="controller-manager" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.013965 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbccfaaf-566f-4e91-84d0-a18b86e848cc" containerName="controller-manager" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.014408 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.065049 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-config\") pod \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.065102 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-client-ca\") pod \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.065130 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-proxy-ca-bundles\") pod \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.065212 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n59m\" (UniqueName: \"kubernetes.io/projected/cbccfaaf-566f-4e91-84d0-a18b86e848cc-kube-api-access-9n59m\") pod \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.065251 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbccfaaf-566f-4e91-84d0-a18b86e848cc-serving-cert\") pod \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\" (UID: \"cbccfaaf-566f-4e91-84d0-a18b86e848cc\") " Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.065951 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-client-ca" (OuterVolumeSpecName: "client-ca") pod "cbccfaaf-566f-4e91-84d0-a18b86e848cc" (UID: "cbccfaaf-566f-4e91-84d0-a18b86e848cc"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.065965 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "cbccfaaf-566f-4e91-84d0-a18b86e848cc" (UID: "cbccfaaf-566f-4e91-84d0-a18b86e848cc"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.066054 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-config" (OuterVolumeSpecName: "config") pod "cbccfaaf-566f-4e91-84d0-a18b86e848cc" (UID: "cbccfaaf-566f-4e91-84d0-a18b86e848cc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.070839 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbccfaaf-566f-4e91-84d0-a18b86e848cc-kube-api-access-9n59m" (OuterVolumeSpecName: "kube-api-access-9n59m") pod "cbccfaaf-566f-4e91-84d0-a18b86e848cc" (UID: "cbccfaaf-566f-4e91-84d0-a18b86e848cc"). InnerVolumeSpecName "kube-api-access-9n59m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.080829 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-68bbc847b9-998jg"] Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.081666 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbccfaaf-566f-4e91-84d0-a18b86e848cc-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cbccfaaf-566f-4e91-84d0-a18b86e848cc" (UID: "cbccfaaf-566f-4e91-84d0-a18b86e848cc"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.166378 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8vgf\" (UniqueName: \"kubernetes.io/projected/4a65970e-4db4-4dff-8a71-28f6936bd13a-kube-api-access-g8vgf\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.166462 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a65970e-4db4-4dff-8a71-28f6936bd13a-config\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.166528 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a65970e-4db4-4dff-8a71-28f6936bd13a-proxy-ca-bundles\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.166567 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a65970e-4db4-4dff-8a71-28f6936bd13a-client-ca\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.166593 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a65970e-4db4-4dff-8a71-28f6936bd13a-serving-cert\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.166631 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.166641 4815 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.166650 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n59m\" (UniqueName: \"kubernetes.io/projected/cbccfaaf-566f-4e91-84d0-a18b86e848cc-kube-api-access-9n59m\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.166658 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbccfaaf-566f-4e91-84d0-a18b86e848cc-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.166669 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbccfaaf-566f-4e91-84d0-a18b86e848cc-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.219656 4815 generic.go:334] "Generic (PLEG): container finished" podID="cbccfaaf-566f-4e91-84d0-a18b86e848cc" containerID="12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005" exitCode=0 Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.219699 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.219701 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" event={"ID":"cbccfaaf-566f-4e91-84d0-a18b86e848cc","Type":"ContainerDied","Data":"12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005"} Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.219818 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5964cbcb45-6ckhp" event={"ID":"cbccfaaf-566f-4e91-84d0-a18b86e848cc","Type":"ContainerDied","Data":"5ca6eb659fab3317012a199b3096494a891e15e1c05a47975acd3732a40fcb3b"} Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.219843 4815 scope.go:117] "RemoveContainer" containerID="12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.235316 4815 scope.go:117] "RemoveContainer" containerID="12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005" Dec 05 09:12:39 crc kubenswrapper[4815]: E1205 09:12:39.235724 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005\": container with ID starting with 12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005 not found: ID does not exist" containerID="12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.235753 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005"} err="failed to get container status \"12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005\": rpc error: code = NotFound desc = could not find container \"12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005\": container with ID starting with 12b1acba064e490154115ce7a35e44b4014284453296ae921b7dbc4dcbd7d005 not found: ID does not exist" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.245915 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5964cbcb45-6ckhp"] Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.249302 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5964cbcb45-6ckhp"] Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.267859 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a65970e-4db4-4dff-8a71-28f6936bd13a-serving-cert\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.267923 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8vgf\" (UniqueName: \"kubernetes.io/projected/4a65970e-4db4-4dff-8a71-28f6936bd13a-kube-api-access-g8vgf\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.267960 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a65970e-4db4-4dff-8a71-28f6936bd13a-config\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.267990 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a65970e-4db4-4dff-8a71-28f6936bd13a-proxy-ca-bundles\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.268029 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a65970e-4db4-4dff-8a71-28f6936bd13a-client-ca\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.269730 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4a65970e-4db4-4dff-8a71-28f6936bd13a-client-ca\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.269730 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4a65970e-4db4-4dff-8a71-28f6936bd13a-proxy-ca-bundles\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.270033 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4a65970e-4db4-4dff-8a71-28f6936bd13a-config\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.282810 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4a65970e-4db4-4dff-8a71-28f6936bd13a-serving-cert\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.289847 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8vgf\" (UniqueName: \"kubernetes.io/projected/4a65970e-4db4-4dff-8a71-28f6936bd13a-kube-api-access-g8vgf\") pod \"controller-manager-68bbc847b9-998jg\" (UID: \"4a65970e-4db4-4dff-8a71-28f6936bd13a\") " pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.332926 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.428119 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbccfaaf-566f-4e91-84d0-a18b86e848cc" path="/var/lib/kubelet/pods/cbccfaaf-566f-4e91-84d0-a18b86e848cc/volumes" Dec 05 09:12:39 crc kubenswrapper[4815]: I1205 09:12:39.725788 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-68bbc847b9-998jg"] Dec 05 09:12:39 crc kubenswrapper[4815]: W1205 09:12:39.731675 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4a65970e_4db4_4dff_8a71_28f6936bd13a.slice/crio-e06910394703a57a7db22279d5243760041b434a2011adb9e5e5561f04950bf7 WatchSource:0}: Error finding container e06910394703a57a7db22279d5243760041b434a2011adb9e5e5561f04950bf7: Status 404 returned error can't find the container with id e06910394703a57a7db22279d5243760041b434a2011adb9e5e5561f04950bf7 Dec 05 09:12:40 crc kubenswrapper[4815]: I1205 09:12:40.226648 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" event={"ID":"4a65970e-4db4-4dff-8a71-28f6936bd13a","Type":"ContainerStarted","Data":"0176579a3c235b674271ed7ed3add75abfa5ab7f71c6b534e7fa4b41bb560bf9"} Dec 05 09:12:40 crc kubenswrapper[4815]: I1205 09:12:40.226695 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" event={"ID":"4a65970e-4db4-4dff-8a71-28f6936bd13a","Type":"ContainerStarted","Data":"e06910394703a57a7db22279d5243760041b434a2011adb9e5e5561f04950bf7"} Dec 05 09:12:40 crc kubenswrapper[4815]: I1205 09:12:40.226916 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:40 crc kubenswrapper[4815]: I1205 09:12:40.234083 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" Dec 05 09:12:40 crc kubenswrapper[4815]: I1205 09:12:40.250530 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-68bbc847b9-998jg" podStartSLOduration=6.250508999 podStartE2EDuration="6.250508999s" podCreationTimestamp="2025-12-05 09:12:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:12:40.248729209 +0000 UTC m=+359.127336056" watchObservedRunningTime="2025-12-05 09:12:40.250508999 +0000 UTC m=+359.129115836" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.556396 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-85r8r"] Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.557748 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.653864 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-85r8r"] Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.750369 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5aef7fd7-1565-4242-a536-b34c07f71443-registry-certificates\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.750561 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5aef7fd7-1565-4242-a536-b34c07f71443-registry-tls\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.750595 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5aef7fd7-1565-4242-a536-b34c07f71443-bound-sa-token\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.750617 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5aef7fd7-1565-4242-a536-b34c07f71443-installation-pull-secrets\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.750651 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.750844 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5aef7fd7-1565-4242-a536-b34c07f71443-ca-trust-extracted\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.750957 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5aef7fd7-1565-4242-a536-b34c07f71443-trusted-ca\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.751004 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr55p\" (UniqueName: \"kubernetes.io/projected/5aef7fd7-1565-4242-a536-b34c07f71443-kube-api-access-wr55p\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.782724 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.852126 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5aef7fd7-1565-4242-a536-b34c07f71443-ca-trust-extracted\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.852221 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5aef7fd7-1565-4242-a536-b34c07f71443-trusted-ca\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.852270 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr55p\" (UniqueName: \"kubernetes.io/projected/5aef7fd7-1565-4242-a536-b34c07f71443-kube-api-access-wr55p\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.852318 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5aef7fd7-1565-4242-a536-b34c07f71443-registry-certificates\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.852348 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5aef7fd7-1565-4242-a536-b34c07f71443-registry-tls\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.852370 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5aef7fd7-1565-4242-a536-b34c07f71443-bound-sa-token\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.852397 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5aef7fd7-1565-4242-a536-b34c07f71443-installation-pull-secrets\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.853028 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/5aef7fd7-1565-4242-a536-b34c07f71443-ca-trust-extracted\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.854318 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/5aef7fd7-1565-4242-a536-b34c07f71443-registry-certificates\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.854365 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/5aef7fd7-1565-4242-a536-b34c07f71443-trusted-ca\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.861791 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/5aef7fd7-1565-4242-a536-b34c07f71443-registry-tls\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.862005 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/5aef7fd7-1565-4242-a536-b34c07f71443-installation-pull-secrets\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.872545 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5aef7fd7-1565-4242-a536-b34c07f71443-bound-sa-token\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.873172 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr55p\" (UniqueName: \"kubernetes.io/projected/5aef7fd7-1565-4242-a536-b34c07f71443-kube-api-access-wr55p\") pod \"image-registry-66df7c8f76-85r8r\" (UID: \"5aef7fd7-1565-4242-a536-b34c07f71443\") " pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:42 crc kubenswrapper[4815]: I1205 09:12:42.873464 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:43 crc kubenswrapper[4815]: I1205 09:12:43.316898 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-85r8r"] Dec 05 09:12:43 crc kubenswrapper[4815]: W1205 09:12:43.321936 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5aef7fd7_1565_4242_a536_b34c07f71443.slice/crio-7af9a09e1d500018cb9a14265329d492c74f8e087a097d4b38a8fd3ed1038423 WatchSource:0}: Error finding container 7af9a09e1d500018cb9a14265329d492c74f8e087a097d4b38a8fd3ed1038423: Status 404 returned error can't find the container with id 7af9a09e1d500018cb9a14265329d492c74f8e087a097d4b38a8fd3ed1038423 Dec 05 09:12:43 crc kubenswrapper[4815]: I1205 09:12:43.892871 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:43 crc kubenswrapper[4815]: I1205 09:12:43.940672 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-t2dbx" Dec 05 09:12:44 crc kubenswrapper[4815]: I1205 09:12:44.250881 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" event={"ID":"5aef7fd7-1565-4242-a536-b34c07f71443","Type":"ContainerStarted","Data":"4070595e5290b924f086eb1c546304ae504ac5acdc35a44aa640f905f5c09d8c"} Dec 05 09:12:44 crc kubenswrapper[4815]: I1205 09:12:44.250918 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" event={"ID":"5aef7fd7-1565-4242-a536-b34c07f71443","Type":"ContainerStarted","Data":"7af9a09e1d500018cb9a14265329d492c74f8e087a097d4b38a8fd3ed1038423"} Dec 05 09:12:44 crc kubenswrapper[4815]: I1205 09:12:44.250949 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:12:44 crc kubenswrapper[4815]: I1205 09:12:44.279975 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" podStartSLOduration=2.279957686 podStartE2EDuration="2.279957686s" podCreationTimestamp="2025-12-05 09:12:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:12:44.274654415 +0000 UTC m=+363.153261252" watchObservedRunningTime="2025-12-05 09:12:44.279957686 +0000 UTC m=+363.158564523" Dec 05 09:12:50 crc kubenswrapper[4815]: I1205 09:12:50.192731 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:12:50 crc kubenswrapper[4815]: I1205 09:12:50.193410 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:12:54 crc kubenswrapper[4815]: I1205 09:12:54.392571 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst"] Dec 05 09:12:54 crc kubenswrapper[4815]: I1205 09:12:54.393167 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" podUID="4b1e66e9-1c8c-4392-9a04-185c83064989" containerName="route-controller-manager" containerID="cri-o://cdafd09012ba11709b36dd28afabbe54f9d0beef3231988f21047706af291b06" gracePeriod=30 Dec 05 09:12:54 crc kubenswrapper[4815]: I1205 09:12:54.482268 4815 patch_prober.go:28] interesting pod/route-controller-manager-5598468cdf-79hst container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" start-of-body= Dec 05 09:12:54 crc kubenswrapper[4815]: I1205 09:12:54.482372 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" podUID="4b1e66e9-1c8c-4392-9a04-185c83064989" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.306004 4815 generic.go:334] "Generic (PLEG): container finished" podID="4b1e66e9-1c8c-4392-9a04-185c83064989" containerID="cdafd09012ba11709b36dd28afabbe54f9d0beef3231988f21047706af291b06" exitCode=0 Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.306046 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" event={"ID":"4b1e66e9-1c8c-4392-9a04-185c83064989","Type":"ContainerDied","Data":"cdafd09012ba11709b36dd28afabbe54f9d0beef3231988f21047706af291b06"} Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.397660 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.545808 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b1e66e9-1c8c-4392-9a04-185c83064989-serving-cert\") pod \"4b1e66e9-1c8c-4392-9a04-185c83064989\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.545928 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-client-ca\") pod \"4b1e66e9-1c8c-4392-9a04-185c83064989\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.545962 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-config\") pod \"4b1e66e9-1c8c-4392-9a04-185c83064989\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.545990 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7fdrg\" (UniqueName: \"kubernetes.io/projected/4b1e66e9-1c8c-4392-9a04-185c83064989-kube-api-access-7fdrg\") pod \"4b1e66e9-1c8c-4392-9a04-185c83064989\" (UID: \"4b1e66e9-1c8c-4392-9a04-185c83064989\") " Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.546886 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-client-ca" (OuterVolumeSpecName: "client-ca") pod "4b1e66e9-1c8c-4392-9a04-185c83064989" (UID: "4b1e66e9-1c8c-4392-9a04-185c83064989"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.546909 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-config" (OuterVolumeSpecName: "config") pod "4b1e66e9-1c8c-4392-9a04-185c83064989" (UID: "4b1e66e9-1c8c-4392-9a04-185c83064989"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.551401 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b1e66e9-1c8c-4392-9a04-185c83064989-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4b1e66e9-1c8c-4392-9a04-185c83064989" (UID: "4b1e66e9-1c8c-4392-9a04-185c83064989"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.552107 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b1e66e9-1c8c-4392-9a04-185c83064989-kube-api-access-7fdrg" (OuterVolumeSpecName: "kube-api-access-7fdrg") pod "4b1e66e9-1c8c-4392-9a04-185c83064989" (UID: "4b1e66e9-1c8c-4392-9a04-185c83064989"). InnerVolumeSpecName "kube-api-access-7fdrg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.647394 4815 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-client-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.647433 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4b1e66e9-1c8c-4392-9a04-185c83064989-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.647443 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7fdrg\" (UniqueName: \"kubernetes.io/projected/4b1e66e9-1c8c-4392-9a04-185c83064989-kube-api-access-7fdrg\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:55 crc kubenswrapper[4815]: I1205 09:12:55.647454 4815 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4b1e66e9-1c8c-4392-9a04-185c83064989-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.111706 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj"] Dec 05 09:12:56 crc kubenswrapper[4815]: E1205 09:12:56.112116 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b1e66e9-1c8c-4392-9a04-185c83064989" containerName="route-controller-manager" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.112151 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b1e66e9-1c8c-4392-9a04-185c83064989" containerName="route-controller-manager" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.112366 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b1e66e9-1c8c-4392-9a04-185c83064989" containerName="route-controller-manager" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.113059 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.117764 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj"] Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.152346 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ab1ff45-e50b-4644-bd45-2436a76dae21-client-ca\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.152649 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ab1ff45-e50b-4644-bd45-2436a76dae21-serving-cert\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.152727 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vdchg\" (UniqueName: \"kubernetes.io/projected/8ab1ff45-e50b-4644-bd45-2436a76dae21-kube-api-access-vdchg\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.152768 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab1ff45-e50b-4644-bd45-2436a76dae21-config\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.253730 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab1ff45-e50b-4644-bd45-2436a76dae21-config\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.253789 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ab1ff45-e50b-4644-bd45-2436a76dae21-client-ca\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.253814 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ab1ff45-e50b-4644-bd45-2436a76dae21-serving-cert\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.253881 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vdchg\" (UniqueName: \"kubernetes.io/projected/8ab1ff45-e50b-4644-bd45-2436a76dae21-kube-api-access-vdchg\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.255442 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8ab1ff45-e50b-4644-bd45-2436a76dae21-client-ca\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.255903 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8ab1ff45-e50b-4644-bd45-2436a76dae21-config\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.262515 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8ab1ff45-e50b-4644-bd45-2436a76dae21-serving-cert\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.270723 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vdchg\" (UniqueName: \"kubernetes.io/projected/8ab1ff45-e50b-4644-bd45-2436a76dae21-kube-api-access-vdchg\") pod \"route-controller-manager-769fd76f7f-bm6zj\" (UID: \"8ab1ff45-e50b-4644-bd45-2436a76dae21\") " pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.313845 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" event={"ID":"4b1e66e9-1c8c-4392-9a04-185c83064989","Type":"ContainerDied","Data":"a60d05731ffdbdb3bc5c90a17ad7b1184116a7dfe4d30d516df6116cb4302ce7"} Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.313909 4815 scope.go:117] "RemoveContainer" containerID="cdafd09012ba11709b36dd28afabbe54f9d0beef3231988f21047706af291b06" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.314057 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.352623 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst"] Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.352684 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5598468cdf-79hst"] Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.448727 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:56 crc kubenswrapper[4815]: I1205 09:12:56.918420 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj"] Dec 05 09:12:57 crc kubenswrapper[4815]: I1205 09:12:57.320109 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" event={"ID":"8ab1ff45-e50b-4644-bd45-2436a76dae21","Type":"ContainerStarted","Data":"bfecbd60e3a662b5d6afecb0e41d521fcd0bf5b9b58e72025c73ddd6cd5d5c8b"} Dec 05 09:12:57 crc kubenswrapper[4815]: I1205 09:12:57.320159 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" event={"ID":"8ab1ff45-e50b-4644-bd45-2436a76dae21","Type":"ContainerStarted","Data":"cbe4c3a67b0529bd34d710a82e42bf345eada064c4d0b4aa6ad5f24806402d7b"} Dec 05 09:12:57 crc kubenswrapper[4815]: I1205 09:12:57.320343 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:12:57 crc kubenswrapper[4815]: I1205 09:12:57.336958 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" podStartSLOduration=3.336938728 podStartE2EDuration="3.336938728s" podCreationTimestamp="2025-12-05 09:12:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:12:57.335380815 +0000 UTC m=+376.213987672" watchObservedRunningTime="2025-12-05 09:12:57.336938728 +0000 UTC m=+376.215545565" Dec 05 09:12:57 crc kubenswrapper[4815]: I1205 09:12:57.425445 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b1e66e9-1c8c-4392-9a04-185c83064989" path="/var/lib/kubelet/pods/4b1e66e9-1c8c-4392-9a04-185c83064989/volumes" Dec 05 09:12:57 crc kubenswrapper[4815]: I1205 09:12:57.527402 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" Dec 05 09:13:02 crc kubenswrapper[4815]: I1205 09:13:02.877262 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-85r8r" Dec 05 09:13:02 crc kubenswrapper[4815]: I1205 09:13:02.938596 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fm9ss"] Dec 05 09:13:20 crc kubenswrapper[4815]: I1205 09:13:20.193030 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:13:20 crc kubenswrapper[4815]: I1205 09:13:20.193640 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:13:27 crc kubenswrapper[4815]: I1205 09:13:27.980701 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" podUID="a6ab936d-3cc8-4c09-869a-50783298666d" containerName="registry" containerID="cri-o://cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e" gracePeriod=30 Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.371248 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.478431 4815 generic.go:334] "Generic (PLEG): container finished" podID="a6ab936d-3cc8-4c09-869a-50783298666d" containerID="cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e" exitCode=0 Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.478481 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" event={"ID":"a6ab936d-3cc8-4c09-869a-50783298666d","Type":"ContainerDied","Data":"cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e"} Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.478530 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" event={"ID":"a6ab936d-3cc8-4c09-869a-50783298666d","Type":"ContainerDied","Data":"cb861f3ef2e5bacf5b59fe8fd46010103a4f29926190e45dd5d90eb20fd3c7e0"} Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.478529 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fm9ss" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.478563 4815 scope.go:117] "RemoveContainer" containerID="cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.497586 4815 scope.go:117] "RemoveContainer" containerID="cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e" Dec 05 09:13:28 crc kubenswrapper[4815]: E1205 09:13:28.498201 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e\": container with ID starting with cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e not found: ID does not exist" containerID="cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.498242 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e"} err="failed to get container status \"cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e\": rpc error: code = NotFound desc = could not find container \"cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e\": container with ID starting with cb1faba23c918654ebb90a5d4c9912a995ffd8212d6a16c12e0060f2c84bec8e not found: ID does not exist" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.509145 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-bound-sa-token\") pod \"a6ab936d-3cc8-4c09-869a-50783298666d\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.509474 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"a6ab936d-3cc8-4c09-869a-50783298666d\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.509550 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a6ab936d-3cc8-4c09-869a-50783298666d-ca-trust-extracted\") pod \"a6ab936d-3cc8-4c09-869a-50783298666d\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.509605 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-trusted-ca\") pod \"a6ab936d-3cc8-4c09-869a-50783298666d\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.510074 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a6ab936d-3cc8-4c09-869a-50783298666d-installation-pull-secrets\") pod \"a6ab936d-3cc8-4c09-869a-50783298666d\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.510224 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-registry-tls\") pod \"a6ab936d-3cc8-4c09-869a-50783298666d\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.510391 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-registry-certificates\") pod \"a6ab936d-3cc8-4c09-869a-50783298666d\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.510524 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mx2jb\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-kube-api-access-mx2jb\") pod \"a6ab936d-3cc8-4c09-869a-50783298666d\" (UID: \"a6ab936d-3cc8-4c09-869a-50783298666d\") " Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.512549 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "a6ab936d-3cc8-4c09-869a-50783298666d" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.513291 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a6ab936d-3cc8-4c09-869a-50783298666d" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.517153 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6ab936d-3cc8-4c09-869a-50783298666d-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "a6ab936d-3cc8-4c09-869a-50783298666d" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.517501 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "a6ab936d-3cc8-4c09-869a-50783298666d" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.518013 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a6ab936d-3cc8-4c09-869a-50783298666d" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.518472 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-kube-api-access-mx2jb" (OuterVolumeSpecName: "kube-api-access-mx2jb") pod "a6ab936d-3cc8-4c09-869a-50783298666d" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d"). InnerVolumeSpecName "kube-api-access-mx2jb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.530658 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6ab936d-3cc8-4c09-869a-50783298666d-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "a6ab936d-3cc8-4c09-869a-50783298666d" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.532204 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "a6ab936d-3cc8-4c09-869a-50783298666d" (UID: "a6ab936d-3cc8-4c09-869a-50783298666d"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.612389 4815 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.612427 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mx2jb\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-kube-api-access-mx2jb\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.612439 4815 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.612450 4815 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/a6ab936d-3cc8-4c09-869a-50783298666d-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.612461 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a6ab936d-3cc8-4c09-869a-50783298666d-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.612475 4815 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/a6ab936d-3cc8-4c09-869a-50783298666d-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.612502 4815 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/a6ab936d-3cc8-4c09-869a-50783298666d-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.806530 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fm9ss"] Dec 05 09:13:28 crc kubenswrapper[4815]: I1205 09:13:28.812033 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fm9ss"] Dec 05 09:13:29 crc kubenswrapper[4815]: I1205 09:13:29.426611 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6ab936d-3cc8-4c09-869a-50783298666d" path="/var/lib/kubelet/pods/a6ab936d-3cc8-4c09-869a-50783298666d/volumes" Dec 05 09:13:50 crc kubenswrapper[4815]: I1205 09:13:50.192090 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:13:50 crc kubenswrapper[4815]: I1205 09:13:50.192595 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:13:50 crc kubenswrapper[4815]: I1205 09:13:50.192661 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:13:50 crc kubenswrapper[4815]: I1205 09:13:50.193113 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1ea7fac19f76fee3446c61cf65967c171412fa2c53983770bc915a74ac684a09"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:13:50 crc kubenswrapper[4815]: I1205 09:13:50.193164 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://1ea7fac19f76fee3446c61cf65967c171412fa2c53983770bc915a74ac684a09" gracePeriod=600 Dec 05 09:13:50 crc kubenswrapper[4815]: I1205 09:13:50.642313 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="1ea7fac19f76fee3446c61cf65967c171412fa2c53983770bc915a74ac684a09" exitCode=0 Dec 05 09:13:50 crc kubenswrapper[4815]: I1205 09:13:50.642691 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"1ea7fac19f76fee3446c61cf65967c171412fa2c53983770bc915a74ac684a09"} Dec 05 09:13:50 crc kubenswrapper[4815]: I1205 09:13:50.642728 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"303d1426a6c74f5c6db6ccfcc51be256883ffe5988a205fb5fa30c6e09f6ea39"} Dec 05 09:13:50 crc kubenswrapper[4815]: I1205 09:13:50.642750 4815 scope.go:117] "RemoveContainer" containerID="d2a1af588b1971e300a3a03173ae2c26508396b379a651f6b9fad4bd03b82be2" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.174833 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd"] Dec 05 09:15:00 crc kubenswrapper[4815]: E1205 09:15:00.175797 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6ab936d-3cc8-4c09-869a-50783298666d" containerName="registry" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.175819 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6ab936d-3cc8-4c09-869a-50783298666d" containerName="registry" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.175986 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6ab936d-3cc8-4c09-869a-50783298666d" containerName="registry" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.176586 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.179210 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.179556 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.192333 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd"] Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.230190 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw5vv\" (UniqueName: \"kubernetes.io/projected/abe6eb7f-9d83-4554-bdb7-1d6e20693873-kube-api-access-hw5vv\") pod \"collect-profiles-29415435-p2dgd\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.230265 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abe6eb7f-9d83-4554-bdb7-1d6e20693873-secret-volume\") pod \"collect-profiles-29415435-p2dgd\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.230352 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abe6eb7f-9d83-4554-bdb7-1d6e20693873-config-volume\") pod \"collect-profiles-29415435-p2dgd\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.331145 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw5vv\" (UniqueName: \"kubernetes.io/projected/abe6eb7f-9d83-4554-bdb7-1d6e20693873-kube-api-access-hw5vv\") pod \"collect-profiles-29415435-p2dgd\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.331241 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abe6eb7f-9d83-4554-bdb7-1d6e20693873-secret-volume\") pod \"collect-profiles-29415435-p2dgd\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.331390 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abe6eb7f-9d83-4554-bdb7-1d6e20693873-config-volume\") pod \"collect-profiles-29415435-p2dgd\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.333197 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abe6eb7f-9d83-4554-bdb7-1d6e20693873-config-volume\") pod \"collect-profiles-29415435-p2dgd\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.340638 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abe6eb7f-9d83-4554-bdb7-1d6e20693873-secret-volume\") pod \"collect-profiles-29415435-p2dgd\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.347972 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw5vv\" (UniqueName: \"kubernetes.io/projected/abe6eb7f-9d83-4554-bdb7-1d6e20693873-kube-api-access-hw5vv\") pod \"collect-profiles-29415435-p2dgd\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.505802 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:00 crc kubenswrapper[4815]: I1205 09:15:00.727533 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd"] Dec 05 09:15:01 crc kubenswrapper[4815]: I1205 09:15:01.040886 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" event={"ID":"abe6eb7f-9d83-4554-bdb7-1d6e20693873","Type":"ContainerStarted","Data":"2b6c0b91412d6ccb39d0c961abd9d16ea90059747f1a0aaa2c4b4c2eff975516"} Dec 05 09:15:02 crc kubenswrapper[4815]: I1205 09:15:02.047883 4815 generic.go:334] "Generic (PLEG): container finished" podID="abe6eb7f-9d83-4554-bdb7-1d6e20693873" containerID="64b87e99d479d12b7384688d39169bbe50246e09c74baa61049e6bf8be19ad2c" exitCode=0 Dec 05 09:15:02 crc kubenswrapper[4815]: I1205 09:15:02.047934 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" event={"ID":"abe6eb7f-9d83-4554-bdb7-1d6e20693873","Type":"ContainerDied","Data":"64b87e99d479d12b7384688d39169bbe50246e09c74baa61049e6bf8be19ad2c"} Dec 05 09:15:03 crc kubenswrapper[4815]: I1205 09:15:03.264987 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:03 crc kubenswrapper[4815]: I1205 09:15:03.272120 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abe6eb7f-9d83-4554-bdb7-1d6e20693873-config-volume\") pod \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " Dec 05 09:15:03 crc kubenswrapper[4815]: I1205 09:15:03.272159 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hw5vv\" (UniqueName: \"kubernetes.io/projected/abe6eb7f-9d83-4554-bdb7-1d6e20693873-kube-api-access-hw5vv\") pod \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " Dec 05 09:15:03 crc kubenswrapper[4815]: I1205 09:15:03.272232 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abe6eb7f-9d83-4554-bdb7-1d6e20693873-secret-volume\") pod \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\" (UID: \"abe6eb7f-9d83-4554-bdb7-1d6e20693873\") " Dec 05 09:15:03 crc kubenswrapper[4815]: I1205 09:15:03.273785 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/abe6eb7f-9d83-4554-bdb7-1d6e20693873-config-volume" (OuterVolumeSpecName: "config-volume") pod "abe6eb7f-9d83-4554-bdb7-1d6e20693873" (UID: "abe6eb7f-9d83-4554-bdb7-1d6e20693873"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:15:03 crc kubenswrapper[4815]: I1205 09:15:03.279648 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abe6eb7f-9d83-4554-bdb7-1d6e20693873-kube-api-access-hw5vv" (OuterVolumeSpecName: "kube-api-access-hw5vv") pod "abe6eb7f-9d83-4554-bdb7-1d6e20693873" (UID: "abe6eb7f-9d83-4554-bdb7-1d6e20693873"). InnerVolumeSpecName "kube-api-access-hw5vv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:15:03 crc kubenswrapper[4815]: I1205 09:15:03.279656 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe6eb7f-9d83-4554-bdb7-1d6e20693873-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "abe6eb7f-9d83-4554-bdb7-1d6e20693873" (UID: "abe6eb7f-9d83-4554-bdb7-1d6e20693873"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:15:03 crc kubenswrapper[4815]: I1205 09:15:03.373284 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/abe6eb7f-9d83-4554-bdb7-1d6e20693873-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:03 crc kubenswrapper[4815]: I1205 09:15:03.373333 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/abe6eb7f-9d83-4554-bdb7-1d6e20693873-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:03 crc kubenswrapper[4815]: I1205 09:15:03.373348 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hw5vv\" (UniqueName: \"kubernetes.io/projected/abe6eb7f-9d83-4554-bdb7-1d6e20693873-kube-api-access-hw5vv\") on node \"crc\" DevicePath \"\"" Dec 05 09:15:04 crc kubenswrapper[4815]: I1205 09:15:04.057868 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" event={"ID":"abe6eb7f-9d83-4554-bdb7-1d6e20693873","Type":"ContainerDied","Data":"2b6c0b91412d6ccb39d0c961abd9d16ea90059747f1a0aaa2c4b4c2eff975516"} Dec 05 09:15:04 crc kubenswrapper[4815]: I1205 09:15:04.057922 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b6c0b91412d6ccb39d0c961abd9d16ea90059747f1a0aaa2c4b4c2eff975516" Dec 05 09:15:04 crc kubenswrapper[4815]: I1205 09:15:04.057924 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd" Dec 05 09:15:50 crc kubenswrapper[4815]: I1205 09:15:50.192702 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:15:50 crc kubenswrapper[4815]: I1205 09:15:50.193455 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:16:20 crc kubenswrapper[4815]: I1205 09:16:20.192794 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:16:20 crc kubenswrapper[4815]: I1205 09:16:20.193558 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:16:50 crc kubenswrapper[4815]: I1205 09:16:50.192155 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:16:50 crc kubenswrapper[4815]: I1205 09:16:50.193058 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:16:50 crc kubenswrapper[4815]: I1205 09:16:50.193232 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:16:50 crc kubenswrapper[4815]: I1205 09:16:50.194248 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"303d1426a6c74f5c6db6ccfcc51be256883ffe5988a205fb5fa30c6e09f6ea39"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:16:50 crc kubenswrapper[4815]: I1205 09:16:50.194356 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://303d1426a6c74f5c6db6ccfcc51be256883ffe5988a205fb5fa30c6e09f6ea39" gracePeriod=600 Dec 05 09:16:51 crc kubenswrapper[4815]: I1205 09:16:51.662108 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="303d1426a6c74f5c6db6ccfcc51be256883ffe5988a205fb5fa30c6e09f6ea39" exitCode=0 Dec 05 09:16:51 crc kubenswrapper[4815]: I1205 09:16:51.662183 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"303d1426a6c74f5c6db6ccfcc51be256883ffe5988a205fb5fa30c6e09f6ea39"} Dec 05 09:16:51 crc kubenswrapper[4815]: I1205 09:16:51.662624 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"0c25c98ba21766bd050ab40ffa71bd173c8e738150c82cbdfb30f46550e857a7"} Dec 05 09:16:51 crc kubenswrapper[4815]: I1205 09:16:51.662653 4815 scope.go:117] "RemoveContainer" containerID="1ea7fac19f76fee3446c61cf65967c171412fa2c53983770bc915a74ac684a09" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.130356 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-zc6mp"] Dec 05 09:18:23 crc kubenswrapper[4815]: E1205 09:18:23.130998 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abe6eb7f-9d83-4554-bdb7-1d6e20693873" containerName="collect-profiles" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.131010 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="abe6eb7f-9d83-4554-bdb7-1d6e20693873" containerName="collect-profiles" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.131102 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="abe6eb7f-9d83-4554-bdb7-1d6e20693873" containerName="collect-profiles" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.131442 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-zc6mp" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.133628 4815 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-cr5jz" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.133691 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.135897 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.141859 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-zc6mp"] Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.184182 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vcstd"] Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.184829 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vcstd" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.186458 4815 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-j5plt" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.190522 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vcstd"] Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.205397 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-sz66k"] Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.206241 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-sz66k" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.210225 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8zv4\" (UniqueName: \"kubernetes.io/projected/8f3da4f1-671e-4911-a481-45824ce433d5-kube-api-access-b8zv4\") pod \"cert-manager-cainjector-7f985d654d-zc6mp\" (UID: \"8f3da4f1-671e-4911-a481-45824ce433d5\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-zc6mp" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.210253 4815 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-qp42r" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.210312 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6n9s\" (UniqueName: \"kubernetes.io/projected/56e3935a-c7dd-4bb2-9dff-deaf3dd7e25b-kube-api-access-m6n9s\") pod \"cert-manager-webhook-5655c58dd6-sz66k\" (UID: \"56e3935a-c7dd-4bb2-9dff-deaf3dd7e25b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-sz66k" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.210367 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwhpm\" (UniqueName: \"kubernetes.io/projected/82c69bfe-d893-41ec-bc5d-d872d51c8c15-kube-api-access-kwhpm\") pod \"cert-manager-5b446d88c5-vcstd\" (UID: \"82c69bfe-d893-41ec-bc5d-d872d51c8c15\") " pod="cert-manager/cert-manager-5b446d88c5-vcstd" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.225670 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-sz66k"] Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.311584 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6n9s\" (UniqueName: \"kubernetes.io/projected/56e3935a-c7dd-4bb2-9dff-deaf3dd7e25b-kube-api-access-m6n9s\") pod \"cert-manager-webhook-5655c58dd6-sz66k\" (UID: \"56e3935a-c7dd-4bb2-9dff-deaf3dd7e25b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-sz66k" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.311656 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwhpm\" (UniqueName: \"kubernetes.io/projected/82c69bfe-d893-41ec-bc5d-d872d51c8c15-kube-api-access-kwhpm\") pod \"cert-manager-5b446d88c5-vcstd\" (UID: \"82c69bfe-d893-41ec-bc5d-d872d51c8c15\") " pod="cert-manager/cert-manager-5b446d88c5-vcstd" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.311697 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8zv4\" (UniqueName: \"kubernetes.io/projected/8f3da4f1-671e-4911-a481-45824ce433d5-kube-api-access-b8zv4\") pod \"cert-manager-cainjector-7f985d654d-zc6mp\" (UID: \"8f3da4f1-671e-4911-a481-45824ce433d5\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-zc6mp" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.329731 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwhpm\" (UniqueName: \"kubernetes.io/projected/82c69bfe-d893-41ec-bc5d-d872d51c8c15-kube-api-access-kwhpm\") pod \"cert-manager-5b446d88c5-vcstd\" (UID: \"82c69bfe-d893-41ec-bc5d-d872d51c8c15\") " pod="cert-manager/cert-manager-5b446d88c5-vcstd" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.330348 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8zv4\" (UniqueName: \"kubernetes.io/projected/8f3da4f1-671e-4911-a481-45824ce433d5-kube-api-access-b8zv4\") pod \"cert-manager-cainjector-7f985d654d-zc6mp\" (UID: \"8f3da4f1-671e-4911-a481-45824ce433d5\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-zc6mp" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.336474 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6n9s\" (UniqueName: \"kubernetes.io/projected/56e3935a-c7dd-4bb2-9dff-deaf3dd7e25b-kube-api-access-m6n9s\") pod \"cert-manager-webhook-5655c58dd6-sz66k\" (UID: \"56e3935a-c7dd-4bb2-9dff-deaf3dd7e25b\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-sz66k" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.448799 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-zc6mp" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.498126 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-vcstd" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.520323 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-sz66k" Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.679503 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-zc6mp"] Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.685808 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.966233 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-sz66k"] Dec 05 09:18:23 crc kubenswrapper[4815]: W1205 09:18:23.969573 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82c69bfe_d893_41ec_bc5d_d872d51c8c15.slice/crio-cc8eaa0655b70585655bfc0c513d4533e263e6819cbd3ffe1d634e9936293e93 WatchSource:0}: Error finding container cc8eaa0655b70585655bfc0c513d4533e263e6819cbd3ffe1d634e9936293e93: Status 404 returned error can't find the container with id cc8eaa0655b70585655bfc0c513d4533e263e6819cbd3ffe1d634e9936293e93 Dec 05 09:18:23 crc kubenswrapper[4815]: I1205 09:18:23.972737 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-vcstd"] Dec 05 09:18:23 crc kubenswrapper[4815]: W1205 09:18:23.981235 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56e3935a_c7dd_4bb2_9dff_deaf3dd7e25b.slice/crio-5e7174d57b03f37d7be6a06113ab4c4a7a9fc0a0b6e695f401f90e1be63ad641 WatchSource:0}: Error finding container 5e7174d57b03f37d7be6a06113ab4c4a7a9fc0a0b6e695f401f90e1be63ad641: Status 404 returned error can't find the container with id 5e7174d57b03f37d7be6a06113ab4c4a7a9fc0a0b6e695f401f90e1be63ad641 Dec 05 09:18:24 crc kubenswrapper[4815]: I1205 09:18:24.197743 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vcstd" event={"ID":"82c69bfe-d893-41ec-bc5d-d872d51c8c15","Type":"ContainerStarted","Data":"cc8eaa0655b70585655bfc0c513d4533e263e6819cbd3ffe1d634e9936293e93"} Dec 05 09:18:24 crc kubenswrapper[4815]: I1205 09:18:24.199037 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-sz66k" event={"ID":"56e3935a-c7dd-4bb2-9dff-deaf3dd7e25b","Type":"ContainerStarted","Data":"5e7174d57b03f37d7be6a06113ab4c4a7a9fc0a0b6e695f401f90e1be63ad641"} Dec 05 09:18:24 crc kubenswrapper[4815]: I1205 09:18:24.200688 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-zc6mp" event={"ID":"8f3da4f1-671e-4911-a481-45824ce433d5","Type":"ContainerStarted","Data":"ecfd2c473f0137756b440c49e5e0a7bb3a7877c61ebc256d73228987bcb422de"} Dec 05 09:18:30 crc kubenswrapper[4815]: I1205 09:18:30.233178 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-sz66k" event={"ID":"56e3935a-c7dd-4bb2-9dff-deaf3dd7e25b","Type":"ContainerStarted","Data":"36a35afe1d71729ad77db717f0893d716612e9955dade438f988cf219d1f4e98"} Dec 05 09:18:30 crc kubenswrapper[4815]: I1205 09:18:30.234820 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-sz66k" Dec 05 09:18:30 crc kubenswrapper[4815]: I1205 09:18:30.236005 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-zc6mp" event={"ID":"8f3da4f1-671e-4911-a481-45824ce433d5","Type":"ContainerStarted","Data":"b4c0729dca136f16816573b9b53ce22b13743925cb744e574fdd690f6334c6a7"} Dec 05 09:18:30 crc kubenswrapper[4815]: I1205 09:18:30.237593 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-vcstd" event={"ID":"82c69bfe-d893-41ec-bc5d-d872d51c8c15","Type":"ContainerStarted","Data":"54589df971754776e393e9d2a2304efd49a5c74cf6feb4fe6aecab169c8efbfc"} Dec 05 09:18:30 crc kubenswrapper[4815]: I1205 09:18:30.280806 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-sz66k" podStartSLOduration=2.280339525 podStartE2EDuration="7.280785761s" podCreationTimestamp="2025-12-05 09:18:23 +0000 UTC" firstStartedPulling="2025-12-05 09:18:23.982879093 +0000 UTC m=+702.861485930" lastFinishedPulling="2025-12-05 09:18:28.983325329 +0000 UTC m=+707.861932166" observedRunningTime="2025-12-05 09:18:30.276892867 +0000 UTC m=+709.155499714" watchObservedRunningTime="2025-12-05 09:18:30.280785761 +0000 UTC m=+709.159392598" Dec 05 09:18:30 crc kubenswrapper[4815]: I1205 09:18:30.299185 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-vcstd" podStartSLOduration=2.291974297 podStartE2EDuration="7.299164654s" podCreationTimestamp="2025-12-05 09:18:23 +0000 UTC" firstStartedPulling="2025-12-05 09:18:23.971773516 +0000 UTC m=+702.850380343" lastFinishedPulling="2025-12-05 09:18:28.978963863 +0000 UTC m=+707.857570700" observedRunningTime="2025-12-05 09:18:30.293445291 +0000 UTC m=+709.172052128" watchObservedRunningTime="2025-12-05 09:18:30.299164654 +0000 UTC m=+709.177771491" Dec 05 09:18:30 crc kubenswrapper[4815]: I1205 09:18:30.312907 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-zc6mp" podStartSLOduration=2.020310283 podStartE2EDuration="7.312881842s" podCreationTimestamp="2025-12-05 09:18:23 +0000 UTC" firstStartedPulling="2025-12-05 09:18:23.685627543 +0000 UTC m=+702.564234380" lastFinishedPulling="2025-12-05 09:18:28.978199102 +0000 UTC m=+707.856805939" observedRunningTime="2025-12-05 09:18:30.307721873 +0000 UTC m=+709.186328740" watchObservedRunningTime="2025-12-05 09:18:30.312881842 +0000 UTC m=+709.191488719" Dec 05 09:18:32 crc kubenswrapper[4815]: I1205 09:18:32.978771 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pzlxw"] Dec 05 09:18:32 crc kubenswrapper[4815]: I1205 09:18:32.979246 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovn-controller" containerID="cri-o://bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9" gracePeriod=30 Dec 05 09:18:32 crc kubenswrapper[4815]: I1205 09:18:32.979303 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="sbdb" containerID="cri-o://242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0" gracePeriod=30 Dec 05 09:18:32 crc kubenswrapper[4815]: I1205 09:18:32.979384 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="nbdb" containerID="cri-o://7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1" gracePeriod=30 Dec 05 09:18:32 crc kubenswrapper[4815]: I1205 09:18:32.979406 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="northd" containerID="cri-o://843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7" gracePeriod=30 Dec 05 09:18:32 crc kubenswrapper[4815]: I1205 09:18:32.979446 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a" gracePeriod=30 Dec 05 09:18:32 crc kubenswrapper[4815]: I1205 09:18:32.979511 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="kube-rbac-proxy-node" containerID="cri-o://5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce" gracePeriod=30 Dec 05 09:18:32 crc kubenswrapper[4815]: I1205 09:18:32.979558 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovn-acl-logging" containerID="cri-o://33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028" gracePeriod=30 Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.053027 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" containerID="cri-o://5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76" gracePeriod=30 Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.257676 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-f9lkk_dff16910-bb36-4748-82bb-0f1d90da05b5/kube-multus/2.log" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.258207 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-f9lkk_dff16910-bb36-4748-82bb-0f1d90da05b5/kube-multus/1.log" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.258249 4815 generic.go:334] "Generic (PLEG): container finished" podID="dff16910-bb36-4748-82bb-0f1d90da05b5" containerID="956f4463e9520b777cb14ec518d485199408c873c9afec4b38e56f0e0ec02fbf" exitCode=2 Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.258312 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-f9lkk" event={"ID":"dff16910-bb36-4748-82bb-0f1d90da05b5","Type":"ContainerDied","Data":"956f4463e9520b777cb14ec518d485199408c873c9afec4b38e56f0e0ec02fbf"} Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.258354 4815 scope.go:117] "RemoveContainer" containerID="93039989c0b31974c991ea71ffbc1cf4039730f40788c516539250a57a95b78b" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.258810 4815 scope.go:117] "RemoveContainer" containerID="956f4463e9520b777cb14ec518d485199408c873c9afec4b38e56f0e0ec02fbf" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.259111 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-f9lkk_openshift-multus(dff16910-bb36-4748-82bb-0f1d90da05b5)\"" pod="openshift-multus/multus-f9lkk" podUID="dff16910-bb36-4748-82bb-0f1d90da05b5" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.278333 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovnkube-controller/3.log" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.280106 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovn-acl-logging/0.log" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.280558 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovn-controller/0.log" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.280916 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76" exitCode=0 Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.280939 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0" exitCode=0 Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.280950 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1" exitCode=0 Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.280958 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a" exitCode=0 Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.280967 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce" exitCode=0 Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.280975 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028" exitCode=143 Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.280984 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9" exitCode=143 Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.280984 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76"} Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.281029 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0"} Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.281044 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1"} Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.281057 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a"} Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.281068 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce"} Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.281079 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028"} Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.281091 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9"} Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.493070 4815 scope.go:117] "RemoveContainer" containerID="c689778e09860f623b6575f8162bbdb2b8637d671697aa03cb63cf00e44b17ca" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.783756 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovn-acl-logging/0.log" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.784288 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovn-controller/0.log" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.784733 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834087 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-7q4vr"] Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.834330 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834347 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.834356 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="kube-rbac-proxy-node" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834365 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="kube-rbac-proxy-node" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.834376 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834384 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.834395 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovn-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834402 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovn-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.834411 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="sbdb" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834419 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="sbdb" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.834432 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834461 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.834475 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834483 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.834512 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="kubecfg-setup" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834520 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="kubecfg-setup" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.834531 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="northd" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834538 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="northd" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.834547 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="nbdb" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834555 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="nbdb" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.834565 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovn-acl-logging" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834573 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovn-acl-logging" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834676 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="northd" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834690 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834702 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="sbdb" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834712 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovn-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834720 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834729 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovn-acl-logging" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834737 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="nbdb" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834747 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="kube-rbac-proxy-ovn-metrics" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834758 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="kube-rbac-proxy-node" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.834767 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.835600 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.835649 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: E1205 09:18:33.835675 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.835684 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.836474 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.836525 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e930cfad-5770-4db6-826b-353554e2f23c" containerName="ovnkube-controller" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.845660 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964081 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-etc-openvswitch\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964184 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-ovn-kubernetes\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964283 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964342 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-systemd-units\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964410 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964375 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-config\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964454 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-log-socket\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964586 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-log-socket" (OuterVolumeSpecName: "log-socket") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964637 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964834 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964896 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.964929 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-bin\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.965107 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-var-lib-cni-networks-ovn-kubernetes\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.965268 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-env-overrides\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.965580 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e930cfad-5770-4db6-826b-353554e2f23c-ovn-node-metrics-cert\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.965705 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-slash\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.965831 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-openvswitch\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.965931 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-netns\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966025 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-systemd\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966127 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-node-log\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966226 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-ovn\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966329 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-script-lib\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966416 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgqrd\" (UniqueName: \"kubernetes.io/projected/e930cfad-5770-4db6-826b-353554e2f23c-kube-api-access-bgqrd\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966534 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-var-lib-openvswitch\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966636 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-netd\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.965150 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.965749 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-slash" (OuterVolumeSpecName: "host-slash") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.965864 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.965973 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.965974 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966158 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-node-log" (OuterVolumeSpecName: "node-log") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966281 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966687 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966796 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966804 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.966900 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.967171 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-kubelet\") pod \"e930cfad-5770-4db6-826b-353554e2f23c\" (UID: \"e930cfad-5770-4db6-826b-353554e2f23c\") " Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.967424 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-cni-bin\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.967539 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c4cb3351-3a1f-4552-a820-841dea326932-ovnkube-config\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.967648 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.967751 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-kubelet\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.967879 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-run-systemd\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.967913 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-run-openvswitch\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.967944 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbqr8\" (UniqueName: \"kubernetes.io/projected/c4cb3351-3a1f-4552-a820-841dea326932-kube-api-access-cbqr8\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.967971 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-slash\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968046 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-cni-netd\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968091 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c4cb3351-3a1f-4552-a820-841dea326932-ovnkube-script-lib\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968122 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-run-ovn\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968155 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c4cb3351-3a1f-4552-a820-841dea326932-ovn-node-metrics-cert\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968179 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-run-ovn-kubernetes\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968222 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-run-netns\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968246 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-systemd-units\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968319 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-etc-openvswitch\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968353 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-var-lib-openvswitch\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968390 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-log-socket\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968414 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-node-log\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968452 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c4cb3351-3a1f-4552-a820-841dea326932-env-overrides\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968669 4815 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968684 4815 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968694 4815 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-node-log\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968702 4815 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968712 4815 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968721 4815 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968729 4815 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968737 4815 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968747 4815 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968756 4815 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968765 4815 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968773 4815 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968780 4815 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-log-socket\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968788 4815 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968796 4815 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968805 4815 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/e930cfad-5770-4db6-826b-353554e2f23c-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.968813 4815 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-host-slash\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.971173 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e930cfad-5770-4db6-826b-353554e2f23c-kube-api-access-bgqrd" (OuterVolumeSpecName: "kube-api-access-bgqrd") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "kube-api-access-bgqrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.971247 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e930cfad-5770-4db6-826b-353554e2f23c-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:18:33 crc kubenswrapper[4815]: I1205 09:18:33.978681 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "e930cfad-5770-4db6-826b-353554e2f23c" (UID: "e930cfad-5770-4db6-826b-353554e2f23c"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069642 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-kubelet\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069699 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069726 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-run-systemd\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069756 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-run-openvswitch\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069775 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbqr8\" (UniqueName: \"kubernetes.io/projected/c4cb3351-3a1f-4552-a820-841dea326932-kube-api-access-cbqr8\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069793 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-slash\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069818 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-cni-netd\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069838 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c4cb3351-3a1f-4552-a820-841dea326932-ovnkube-script-lib\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069858 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-run-ovn\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069871 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c4cb3351-3a1f-4552-a820-841dea326932-ovn-node-metrics-cert\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069886 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-run-ovn-kubernetes\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069905 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-run-netns\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069921 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-systemd-units\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069943 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-etc-openvswitch\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069960 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-var-lib-openvswitch\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069977 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-log-socket\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.069994 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-node-log\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.070015 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c4cb3351-3a1f-4552-a820-841dea326932-env-overrides\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.070072 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-cni-bin\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.070088 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c4cb3351-3a1f-4552-a820-841dea326932-ovnkube-config\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.070126 4815 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/e930cfad-5770-4db6-826b-353554e2f23c-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.070380 4815 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/e930cfad-5770-4db6-826b-353554e2f23c-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.070389 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgqrd\" (UniqueName: \"kubernetes.io/projected/e930cfad-5770-4db6-826b-353554e2f23c-kube-api-access-bgqrd\") on node \"crc\" DevicePath \"\"" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.070972 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-kubelet\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071063 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071096 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-run-systemd\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071126 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-run-openvswitch\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071157 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-var-lib-openvswitch\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071187 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-run-ovn-kubernetes\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071215 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-run-netns\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071241 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-systemd-units\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071268 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-etc-openvswitch\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071299 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-cni-netd\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071336 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-slash\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071558 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c4cb3351-3a1f-4552-a820-841dea326932-ovnkube-config\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071610 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-log-socket\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071636 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-node-log\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071975 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c4cb3351-3a1f-4552-a820-841dea326932-env-overrides\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.071995 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c4cb3351-3a1f-4552-a820-841dea326932-ovnkube-script-lib\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.072011 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-host-cni-bin\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.072042 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c4cb3351-3a1f-4552-a820-841dea326932-run-ovn\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.075386 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c4cb3351-3a1f-4552-a820-841dea326932-ovn-node-metrics-cert\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.086715 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbqr8\" (UniqueName: \"kubernetes.io/projected/c4cb3351-3a1f-4552-a820-841dea326932-kube-api-access-cbqr8\") pod \"ovnkube-node-7q4vr\" (UID: \"c4cb3351-3a1f-4552-a820-841dea326932\") " pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.174871 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.289260 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-f9lkk_dff16910-bb36-4748-82bb-0f1d90da05b5/kube-multus/2.log" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.293206 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovn-acl-logging/0.log" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.294000 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-pzlxw_e930cfad-5770-4db6-826b-353554e2f23c/ovn-controller/0.log" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.294732 4815 generic.go:334] "Generic (PLEG): container finished" podID="e930cfad-5770-4db6-826b-353554e2f23c" containerID="843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7" exitCode=0 Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.294816 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7"} Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.294846 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" event={"ID":"e930cfad-5770-4db6-826b-353554e2f23c","Type":"ContainerDied","Data":"9ce7716475da15fbe3f700791633122ab8e9d196e7a038a6ab2483390928fe29"} Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.294865 4815 scope.go:117] "RemoveContainer" containerID="5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.295235 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pzlxw" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.297058 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" event={"ID":"c4cb3351-3a1f-4552-a820-841dea326932","Type":"ContainerStarted","Data":"0070f8b88cbca15f60b2f99c2f5cc5aa8bac5a8cbe4e1133b6acaae33acbf926"} Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.314853 4815 scope.go:117] "RemoveContainer" containerID="242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.371731 4815 scope.go:117] "RemoveContainer" containerID="7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.381006 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pzlxw"] Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.389815 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pzlxw"] Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.392181 4815 scope.go:117] "RemoveContainer" containerID="843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.406379 4815 scope.go:117] "RemoveContainer" containerID="addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.420227 4815 scope.go:117] "RemoveContainer" containerID="5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.433164 4815 scope.go:117] "RemoveContainer" containerID="33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.446130 4815 scope.go:117] "RemoveContainer" containerID="bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.458877 4815 scope.go:117] "RemoveContainer" containerID="57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.472151 4815 scope.go:117] "RemoveContainer" containerID="5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76" Dec 05 09:18:34 crc kubenswrapper[4815]: E1205 09:18:34.472665 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76\": container with ID starting with 5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76 not found: ID does not exist" containerID="5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.472693 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76"} err="failed to get container status \"5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76\": rpc error: code = NotFound desc = could not find container \"5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76\": container with ID starting with 5119a6163fd312130e7942dcf0622756c9539a173909f426932af5957ffebd76 not found: ID does not exist" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.472713 4815 scope.go:117] "RemoveContainer" containerID="242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0" Dec 05 09:18:34 crc kubenswrapper[4815]: E1205 09:18:34.473043 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\": container with ID starting with 242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0 not found: ID does not exist" containerID="242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.473072 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0"} err="failed to get container status \"242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\": rpc error: code = NotFound desc = could not find container \"242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0\": container with ID starting with 242bda4940b0821d0487d38f5a365167fad2590eae7a5932783d6360de03e7b0 not found: ID does not exist" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.473089 4815 scope.go:117] "RemoveContainer" containerID="7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1" Dec 05 09:18:34 crc kubenswrapper[4815]: E1205 09:18:34.473679 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\": container with ID starting with 7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1 not found: ID does not exist" containerID="7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.473697 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1"} err="failed to get container status \"7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\": rpc error: code = NotFound desc = could not find container \"7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1\": container with ID starting with 7bec8bbade1a21881c4bf50ec02105a4b07672b3b42ca1eeece37f808c3a60e1 not found: ID does not exist" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.473726 4815 scope.go:117] "RemoveContainer" containerID="843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7" Dec 05 09:18:34 crc kubenswrapper[4815]: E1205 09:18:34.473954 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\": container with ID starting with 843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7 not found: ID does not exist" containerID="843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.474011 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7"} err="failed to get container status \"843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\": rpc error: code = NotFound desc = could not find container \"843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7\": container with ID starting with 843deccb3fd443eb6d025bd4538ce8131c8cc86f4fba502078d2e905bf5fb5c7 not found: ID does not exist" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.474035 4815 scope.go:117] "RemoveContainer" containerID="addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a" Dec 05 09:18:34 crc kubenswrapper[4815]: E1205 09:18:34.474379 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\": container with ID starting with addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a not found: ID does not exist" containerID="addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.474397 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a"} err="failed to get container status \"addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\": rpc error: code = NotFound desc = could not find container \"addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a\": container with ID starting with addbf028b5a61fc6b68fffada78235157dcc47093a72e337891e6cd848d4dc5a not found: ID does not exist" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.474410 4815 scope.go:117] "RemoveContainer" containerID="5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce" Dec 05 09:18:34 crc kubenswrapper[4815]: E1205 09:18:34.474678 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\": container with ID starting with 5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce not found: ID does not exist" containerID="5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.474694 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce"} err="failed to get container status \"5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\": rpc error: code = NotFound desc = could not find container \"5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce\": container with ID starting with 5fedde6a0fea56eadd728f5699ff6163009f7f8b2493cf02c4185e83e979fdce not found: ID does not exist" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.474705 4815 scope.go:117] "RemoveContainer" containerID="33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028" Dec 05 09:18:34 crc kubenswrapper[4815]: E1205 09:18:34.474946 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\": container with ID starting with 33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028 not found: ID does not exist" containerID="33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.474975 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028"} err="failed to get container status \"33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\": rpc error: code = NotFound desc = could not find container \"33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028\": container with ID starting with 33b4f3d6d50c1196d1c40ce973d794884a6288f9d3247bd0964cb12c54956028 not found: ID does not exist" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.474986 4815 scope.go:117] "RemoveContainer" containerID="bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9" Dec 05 09:18:34 crc kubenswrapper[4815]: E1205 09:18:34.475210 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\": container with ID starting with bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9 not found: ID does not exist" containerID="bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.475223 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9"} err="failed to get container status \"bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\": rpc error: code = NotFound desc = could not find container \"bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9\": container with ID starting with bf9a47c81ae00fdff1463a39488ff598aee9170504d93c3e23f21319ab946ea9 not found: ID does not exist" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.475233 4815 scope.go:117] "RemoveContainer" containerID="57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a" Dec 05 09:18:34 crc kubenswrapper[4815]: E1205 09:18:34.475458 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\": container with ID starting with 57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a not found: ID does not exist" containerID="57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a" Dec 05 09:18:34 crc kubenswrapper[4815]: I1205 09:18:34.475477 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a"} err="failed to get container status \"57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\": rpc error: code = NotFound desc = could not find container \"57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a\": container with ID starting with 57297681fc6a3c0d8b4fd89297c21ab8b4f8b130ef49e870f3b91a9a1bb4c48a not found: ID does not exist" Dec 05 09:18:35 crc kubenswrapper[4815]: I1205 09:18:35.303853 4815 generic.go:334] "Generic (PLEG): container finished" podID="c4cb3351-3a1f-4552-a820-841dea326932" containerID="9cf05b839080c2df13862a37b186d068aa0817ae9b69dbd00ff2933cfea46ccd" exitCode=0 Dec 05 09:18:35 crc kubenswrapper[4815]: I1205 09:18:35.303932 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" event={"ID":"c4cb3351-3a1f-4552-a820-841dea326932","Type":"ContainerDied","Data":"9cf05b839080c2df13862a37b186d068aa0817ae9b69dbd00ff2933cfea46ccd"} Dec 05 09:18:35 crc kubenswrapper[4815]: I1205 09:18:35.425521 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e930cfad-5770-4db6-826b-353554e2f23c" path="/var/lib/kubelet/pods/e930cfad-5770-4db6-826b-353554e2f23c/volumes" Dec 05 09:18:36 crc kubenswrapper[4815]: I1205 09:18:36.316789 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" event={"ID":"c4cb3351-3a1f-4552-a820-841dea326932","Type":"ContainerStarted","Data":"ff8ab13463c46f07383bb88c0d3aa727c8bafb834c952daea25adaa5051f8f9e"} Dec 05 09:18:36 crc kubenswrapper[4815]: I1205 09:18:36.317989 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" event={"ID":"c4cb3351-3a1f-4552-a820-841dea326932","Type":"ContainerStarted","Data":"e51a7012a421125b36133331b284ed21beccb3cb2641a7907b36e6765967a07a"} Dec 05 09:18:36 crc kubenswrapper[4815]: I1205 09:18:36.318056 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" event={"ID":"c4cb3351-3a1f-4552-a820-841dea326932","Type":"ContainerStarted","Data":"34b2b91c6377e1472fb1172a18de77cf6fc14eac926b53bffa5c5f4fcd7be042"} Dec 05 09:18:36 crc kubenswrapper[4815]: I1205 09:18:36.318160 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" event={"ID":"c4cb3351-3a1f-4552-a820-841dea326932","Type":"ContainerStarted","Data":"d3579e16143ce4ccb51b9032dff0a1905ec577e5c46a61d7b8271401151c043b"} Dec 05 09:18:36 crc kubenswrapper[4815]: I1205 09:18:36.318223 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" event={"ID":"c4cb3351-3a1f-4552-a820-841dea326932","Type":"ContainerStarted","Data":"a6299fadc121fa8855258204b8266bdc5f98bc11ac5ebfb9a28ea9e1058fb346"} Dec 05 09:18:36 crc kubenswrapper[4815]: I1205 09:18:36.318358 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" event={"ID":"c4cb3351-3a1f-4552-a820-841dea326932","Type":"ContainerStarted","Data":"3cfa6964470c14cf68372923306ce6e834f11d98f0a96030b1c66f43e57c515d"} Dec 05 09:18:38 crc kubenswrapper[4815]: I1205 09:18:38.522826 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-sz66k" Dec 05 09:18:40 crc kubenswrapper[4815]: I1205 09:18:40.339516 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" event={"ID":"c4cb3351-3a1f-4552-a820-841dea326932","Type":"ContainerStarted","Data":"5e146a86a630e96fd7966c679766f4bd2db114a5ebc8faf09b4f51f967a729b3"} Dec 05 09:18:43 crc kubenswrapper[4815]: I1205 09:18:43.359474 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" event={"ID":"c4cb3351-3a1f-4552-a820-841dea326932","Type":"ContainerStarted","Data":"0fea74c224b435661a774a884217be37450463002a01e49a11ad0e7fdb44fba3"} Dec 05 09:18:43 crc kubenswrapper[4815]: I1205 09:18:43.360068 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:43 crc kubenswrapper[4815]: I1205 09:18:43.360087 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:43 crc kubenswrapper[4815]: I1205 09:18:43.360097 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:43 crc kubenswrapper[4815]: I1205 09:18:43.385394 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:43 crc kubenswrapper[4815]: I1205 09:18:43.387671 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" podStartSLOduration=10.387658583 podStartE2EDuration="10.387658583s" podCreationTimestamp="2025-12-05 09:18:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:18:43.385828123 +0000 UTC m=+722.264434960" watchObservedRunningTime="2025-12-05 09:18:43.387658583 +0000 UTC m=+722.266265420" Dec 05 09:18:43 crc kubenswrapper[4815]: I1205 09:18:43.393434 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:18:48 crc kubenswrapper[4815]: I1205 09:18:48.418550 4815 scope.go:117] "RemoveContainer" containerID="956f4463e9520b777cb14ec518d485199408c873c9afec4b38e56f0e0ec02fbf" Dec 05 09:18:49 crc kubenswrapper[4815]: I1205 09:18:49.393199 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-f9lkk_dff16910-bb36-4748-82bb-0f1d90da05b5/kube-multus/2.log" Dec 05 09:18:49 crc kubenswrapper[4815]: I1205 09:18:49.393269 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-f9lkk" event={"ID":"dff16910-bb36-4748-82bb-0f1d90da05b5","Type":"ContainerStarted","Data":"31705d0ae08413ec842d9a8b4f7a445900615c15fcf758300d1094d0bc32f7a0"} Dec 05 09:18:50 crc kubenswrapper[4815]: I1205 09:18:50.192997 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:18:50 crc kubenswrapper[4815]: I1205 09:18:50.193105 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:19:04 crc kubenswrapper[4815]: I1205 09:19:04.202901 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" Dec 05 09:19:20 crc kubenswrapper[4815]: I1205 09:19:20.192851 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:19:20 crc kubenswrapper[4815]: I1205 09:19:20.193340 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.794096 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp"] Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.796776 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.801470 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.805969 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp"] Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.860905 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v57df\" (UniqueName: \"kubernetes.io/projected/f6f8e345-b259-4c23-a98a-22ff6454d46d-kube-api-access-v57df\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.861186 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.861280 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.962105 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.962173 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v57df\" (UniqueName: \"kubernetes.io/projected/f6f8e345-b259-4c23-a98a-22ff6454d46d-kube-api-access-v57df\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.962205 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.962737 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.962762 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:32 crc kubenswrapper[4815]: I1205 09:19:32.986546 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v57df\" (UniqueName: \"kubernetes.io/projected/f6f8e345-b259-4c23-a98a-22ff6454d46d-kube-api-access-v57df\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:33 crc kubenswrapper[4815]: I1205 09:19:33.116176 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:33 crc kubenswrapper[4815]: I1205 09:19:33.469242 4815 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 05 09:19:33 crc kubenswrapper[4815]: I1205 09:19:33.550311 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp"] Dec 05 09:19:33 crc kubenswrapper[4815]: I1205 09:19:33.637471 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" event={"ID":"f6f8e345-b259-4c23-a98a-22ff6454d46d","Type":"ContainerStarted","Data":"ccb4ee05adf5c7d89b43c4bb56cbe774be3fe1f028a7b984f9c231af20c89ed2"} Dec 05 09:19:34 crc kubenswrapper[4815]: I1205 09:19:34.643695 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" event={"ID":"f6f8e345-b259-4c23-a98a-22ff6454d46d","Type":"ContainerStarted","Data":"c99720737172d2fee8e9951341bf810775f6ac5ff336954cd2d280a29253a7b0"} Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.146584 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-7w9wt"] Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.147865 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.167856 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7w9wt"] Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.290058 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mhz5\" (UniqueName: \"kubernetes.io/projected/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-kube-api-access-7mhz5\") pod \"redhat-operators-7w9wt\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.290432 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-catalog-content\") pod \"redhat-operators-7w9wt\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.290581 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-utilities\") pod \"redhat-operators-7w9wt\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.392194 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-catalog-content\") pod \"redhat-operators-7w9wt\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.392664 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-utilities\") pod \"redhat-operators-7w9wt\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.392932 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-catalog-content\") pod \"redhat-operators-7w9wt\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.393264 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-utilities\") pod \"redhat-operators-7w9wt\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.393471 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mhz5\" (UniqueName: \"kubernetes.io/projected/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-kube-api-access-7mhz5\") pod \"redhat-operators-7w9wt\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.414472 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mhz5\" (UniqueName: \"kubernetes.io/projected/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-kube-api-access-7mhz5\") pod \"redhat-operators-7w9wt\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.466566 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.650571 4815 generic.go:334] "Generic (PLEG): container finished" podID="f6f8e345-b259-4c23-a98a-22ff6454d46d" containerID="c99720737172d2fee8e9951341bf810775f6ac5ff336954cd2d280a29253a7b0" exitCode=0 Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.650693 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" event={"ID":"f6f8e345-b259-4c23-a98a-22ff6454d46d","Type":"ContainerDied","Data":"c99720737172d2fee8e9951341bf810775f6ac5ff336954cd2d280a29253a7b0"} Dec 05 09:19:35 crc kubenswrapper[4815]: I1205 09:19:35.727915 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-7w9wt"] Dec 05 09:19:36 crc kubenswrapper[4815]: I1205 09:19:36.659287 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w9wt" event={"ID":"732d7c70-b825-4c00-b7d8-d5f6327ebe1e","Type":"ContainerStarted","Data":"b61d7f61ec758817b835529e9cac4361f3a2f291daf55fb71b1d69754ed5273e"} Dec 05 09:19:37 crc kubenswrapper[4815]: I1205 09:19:37.669076 4815 generic.go:334] "Generic (PLEG): container finished" podID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerID="9731ad28197a6e949ed2eda7641b4faad7859379ea53d8968c1e8bf29969a0d4" exitCode=0 Dec 05 09:19:37 crc kubenswrapper[4815]: I1205 09:19:37.669140 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w9wt" event={"ID":"732d7c70-b825-4c00-b7d8-d5f6327ebe1e","Type":"ContainerDied","Data":"9731ad28197a6e949ed2eda7641b4faad7859379ea53d8968c1e8bf29969a0d4"} Dec 05 09:19:39 crc kubenswrapper[4815]: I1205 09:19:39.686556 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" event={"ID":"f6f8e345-b259-4c23-a98a-22ff6454d46d","Type":"ContainerStarted","Data":"b4aeebe3efc5b52dbb5f17e4d97a16da9f49d4791789b192cea1c7ce3a9c81a1"} Dec 05 09:19:39 crc kubenswrapper[4815]: I1205 09:19:39.688352 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w9wt" event={"ID":"732d7c70-b825-4c00-b7d8-d5f6327ebe1e","Type":"ContainerStarted","Data":"b3a5d4904edfb5cbd50960e7b33811622311622c3bd74e089e33d43d1f3250c1"} Dec 05 09:19:40 crc kubenswrapper[4815]: I1205 09:19:40.708999 4815 generic.go:334] "Generic (PLEG): container finished" podID="f6f8e345-b259-4c23-a98a-22ff6454d46d" containerID="b4aeebe3efc5b52dbb5f17e4d97a16da9f49d4791789b192cea1c7ce3a9c81a1" exitCode=0 Dec 05 09:19:40 crc kubenswrapper[4815]: I1205 09:19:40.709071 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" event={"ID":"f6f8e345-b259-4c23-a98a-22ff6454d46d","Type":"ContainerDied","Data":"b4aeebe3efc5b52dbb5f17e4d97a16da9f49d4791789b192cea1c7ce3a9c81a1"} Dec 05 09:19:42 crc kubenswrapper[4815]: I1205 09:19:42.722351 4815 generic.go:334] "Generic (PLEG): container finished" podID="f6f8e345-b259-4c23-a98a-22ff6454d46d" containerID="6b060295a0355457d21aab04296259a848a31ec160331af484d0204eeebf9702" exitCode=0 Dec 05 09:19:42 crc kubenswrapper[4815]: I1205 09:19:42.722411 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" event={"ID":"f6f8e345-b259-4c23-a98a-22ff6454d46d","Type":"ContainerDied","Data":"6b060295a0355457d21aab04296259a848a31ec160331af484d0204eeebf9702"} Dec 05 09:19:42 crc kubenswrapper[4815]: I1205 09:19:42.724894 4815 generic.go:334] "Generic (PLEG): container finished" podID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerID="b3a5d4904edfb5cbd50960e7b33811622311622c3bd74e089e33d43d1f3250c1" exitCode=0 Dec 05 09:19:42 crc kubenswrapper[4815]: I1205 09:19:42.724961 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w9wt" event={"ID":"732d7c70-b825-4c00-b7d8-d5f6327ebe1e","Type":"ContainerDied","Data":"b3a5d4904edfb5cbd50960e7b33811622311622c3bd74e089e33d43d1f3250c1"} Dec 05 09:19:43 crc kubenswrapper[4815]: I1205 09:19:43.732089 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w9wt" event={"ID":"732d7c70-b825-4c00-b7d8-d5f6327ebe1e","Type":"ContainerStarted","Data":"cff9713891780dd72cf0deae9b26ba0cd1ba06635e6b398f226c83555a65a600"} Dec 05 09:19:43 crc kubenswrapper[4815]: I1205 09:19:43.759014 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-7w9wt" podStartSLOduration=3.2696068 podStartE2EDuration="8.758993235s" podCreationTimestamp="2025-12-05 09:19:35 +0000 UTC" firstStartedPulling="2025-12-05 09:19:37.674119627 +0000 UTC m=+776.552726494" lastFinishedPulling="2025-12-05 09:19:43.163506092 +0000 UTC m=+782.042112929" observedRunningTime="2025-12-05 09:19:43.755849279 +0000 UTC m=+782.634456136" watchObservedRunningTime="2025-12-05 09:19:43.758993235 +0000 UTC m=+782.637600072" Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.100023 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.256900 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-util\") pod \"f6f8e345-b259-4c23-a98a-22ff6454d46d\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.257005 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v57df\" (UniqueName: \"kubernetes.io/projected/f6f8e345-b259-4c23-a98a-22ff6454d46d-kube-api-access-v57df\") pod \"f6f8e345-b259-4c23-a98a-22ff6454d46d\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.257028 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-bundle\") pod \"f6f8e345-b259-4c23-a98a-22ff6454d46d\" (UID: \"f6f8e345-b259-4c23-a98a-22ff6454d46d\") " Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.257566 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-bundle" (OuterVolumeSpecName: "bundle") pod "f6f8e345-b259-4c23-a98a-22ff6454d46d" (UID: "f6f8e345-b259-4c23-a98a-22ff6454d46d"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.265478 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6f8e345-b259-4c23-a98a-22ff6454d46d-kube-api-access-v57df" (OuterVolumeSpecName: "kube-api-access-v57df") pod "f6f8e345-b259-4c23-a98a-22ff6454d46d" (UID: "f6f8e345-b259-4c23-a98a-22ff6454d46d"). InnerVolumeSpecName "kube-api-access-v57df". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.267572 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-util" (OuterVolumeSpecName: "util") pod "f6f8e345-b259-4c23-a98a-22ff6454d46d" (UID: "f6f8e345-b259-4c23-a98a-22ff6454d46d"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.358091 4815 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-util\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.358146 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v57df\" (UniqueName: \"kubernetes.io/projected/f6f8e345-b259-4c23-a98a-22ff6454d46d-kube-api-access-v57df\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.358164 4815 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f6f8e345-b259-4c23-a98a-22ff6454d46d-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.739940 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" event={"ID":"f6f8e345-b259-4c23-a98a-22ff6454d46d","Type":"ContainerDied","Data":"ccb4ee05adf5c7d89b43c4bb56cbe774be3fe1f028a7b984f9c231af20c89ed2"} Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.739978 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ccb4ee05adf5c7d89b43c4bb56cbe774be3fe1f028a7b984f9c231af20c89ed2" Dec 05 09:19:44 crc kubenswrapper[4815]: I1205 09:19:44.739988 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp" Dec 05 09:19:45 crc kubenswrapper[4815]: I1205 09:19:45.467423 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:45 crc kubenswrapper[4815]: I1205 09:19:45.467525 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.114858 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92"] Dec 05 09:19:46 crc kubenswrapper[4815]: E1205 09:19:46.115295 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f8e345-b259-4c23-a98a-22ff6454d46d" containerName="pull" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.115307 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f8e345-b259-4c23-a98a-22ff6454d46d" containerName="pull" Dec 05 09:19:46 crc kubenswrapper[4815]: E1205 09:19:46.115316 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f8e345-b259-4c23-a98a-22ff6454d46d" containerName="util" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.115324 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f8e345-b259-4c23-a98a-22ff6454d46d" containerName="util" Dec 05 09:19:46 crc kubenswrapper[4815]: E1205 09:19:46.115335 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f8e345-b259-4c23-a98a-22ff6454d46d" containerName="extract" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.115341 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f8e345-b259-4c23-a98a-22ff6454d46d" containerName="extract" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.115445 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6f8e345-b259-4c23-a98a-22ff6454d46d" containerName="extract" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.115811 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.120122 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-2ln8r" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.120615 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.130541 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.134150 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92"] Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.188404 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4ldm\" (UniqueName: \"kubernetes.io/projected/7e7d72c9-18b4-4716-aabb-2c55c7070020-kube-api-access-r4ldm\") pod \"nmstate-operator-5b5b58f5c8-82r92\" (UID: \"7e7d72c9-18b4-4716-aabb-2c55c7070020\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.290456 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4ldm\" (UniqueName: \"kubernetes.io/projected/7e7d72c9-18b4-4716-aabb-2c55c7070020-kube-api-access-r4ldm\") pod \"nmstate-operator-5b5b58f5c8-82r92\" (UID: \"7e7d72c9-18b4-4716-aabb-2c55c7070020\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.312455 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4ldm\" (UniqueName: \"kubernetes.io/projected/7e7d72c9-18b4-4716-aabb-2c55c7070020-kube-api-access-r4ldm\") pod \"nmstate-operator-5b5b58f5c8-82r92\" (UID: \"7e7d72c9-18b4-4716-aabb-2c55c7070020\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.429160 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92" Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.526781 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-7w9wt" podUID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerName="registry-server" probeResult="failure" output=< Dec 05 09:19:46 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 09:19:46 crc kubenswrapper[4815]: > Dec 05 09:19:46 crc kubenswrapper[4815]: I1205 09:19:46.776547 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92"] Dec 05 09:19:47 crc kubenswrapper[4815]: I1205 09:19:47.758875 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92" event={"ID":"7e7d72c9-18b4-4716-aabb-2c55c7070020","Type":"ContainerStarted","Data":"2eadea6cef5dfba1dd828eec40734ba4033822b7d6d55b13710b17d4e96ccfa3"} Dec 05 09:19:50 crc kubenswrapper[4815]: I1205 09:19:50.193243 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:19:50 crc kubenswrapper[4815]: I1205 09:19:50.193600 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:19:50 crc kubenswrapper[4815]: I1205 09:19:50.193648 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:19:50 crc kubenswrapper[4815]: I1205 09:19:50.194199 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0c25c98ba21766bd050ab40ffa71bd173c8e738150c82cbdfb30f46550e857a7"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:19:50 crc kubenswrapper[4815]: I1205 09:19:50.194244 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://0c25c98ba21766bd050ab40ffa71bd173c8e738150c82cbdfb30f46550e857a7" gracePeriod=600 Dec 05 09:19:55 crc kubenswrapper[4815]: I1205 09:19:55.508647 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:55 crc kubenswrapper[4815]: I1205 09:19:55.553839 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:19:55 crc kubenswrapper[4815]: I1205 09:19:55.744531 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7w9wt"] Dec 05 09:19:57 crc kubenswrapper[4815]: I1205 09:19:57.506981 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="0c25c98ba21766bd050ab40ffa71bd173c8e738150c82cbdfb30f46550e857a7" exitCode=0 Dec 05 09:19:57 crc kubenswrapper[4815]: I1205 09:19:57.507082 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"0c25c98ba21766bd050ab40ffa71bd173c8e738150c82cbdfb30f46550e857a7"} Dec 05 09:19:57 crc kubenswrapper[4815]: I1205 09:19:57.507628 4815 scope.go:117] "RemoveContainer" containerID="303d1426a6c74f5c6db6ccfcc51be256883ffe5988a205fb5fa30c6e09f6ea39" Dec 05 09:19:59 crc kubenswrapper[4815]: I1205 09:19:59.522600 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-7w9wt" podUID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerName="registry-server" containerID="cri-o://cff9713891780dd72cf0deae9b26ba0cd1ba06635e6b398f226c83555a65a600" gracePeriod=2 Dec 05 09:20:01 crc kubenswrapper[4815]: E1205 09:20:01.989523 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage22533152/1\": happened during read: context canceled" image="registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:dd89e08ed6257597e99b1243839d5c76e6bad72fe9e168c0eba5ce9c449189cf" Dec 05 09:20:01 crc kubenswrapper[4815]: E1205 09:20:01.989948 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nmstate-operator,Image:registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:dd89e08ed6257597e99b1243839d5c76e6bad72fe9e168c0eba5ce9c449189cf,Command:[manager],Args:[--zap-time-encoding=iso8601],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:WATCH_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.annotations['olm.targetNamespaces'],},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:OPERATOR_NAME,Value:kubernetes-nmstate-operator,ValueFrom:nil,},EnvVar{Name:ENABLE_PROFILER,Value:False,ValueFrom:nil,},EnvVar{Name:PROFILER_PORT,Value:6060,ValueFrom:nil,},EnvVar{Name:RUN_OPERATOR,Value:,ValueFrom:nil,},EnvVar{Name:HANDLER_IMAGE,Value:registry.redhat.io/openshift4/ose-kubernetes-nmstate-handler-rhel9@sha256:8045b3d5059cc81bf37964d359055dea9e4915c83f3eec4f800d5ce294c06f97,ValueFrom:nil,},EnvVar{Name:PLUGIN_IMAGE,Value:registry.redhat.io/openshift4/nmstate-console-plugin-rhel9@sha256:10fe26b1ef17d6fa13d22976b553b935f1cc14e74b8dd14a31306554aff7c513,ValueFrom:nil,},EnvVar{Name:HANDLER_IMAGE_PULL_POLICY,Value:Always,ValueFrom:nil,},EnvVar{Name:HANDLER_NAMESPACE,Value:openshift-nmstate,ValueFrom:nil,},EnvVar{Name:MONITORING_NAMESPACE,Value:openshift-monitoring,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:registry.redhat.io/openshift4/ose-kube-rbac-proxy-rhel9@sha256:23ad174e653d608ec2285f670d8669dbe8bb433f7c215bdb59f5c6ac6ad1bcc9,ValueFrom:nil,},EnvVar{Name:OPERATOR_CONDITION_NAME,Value:kubernetes-nmstate-operator.4.18.0-202511191213,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{60 -3} {} 60m DecimalSI},memory: {{31457280 0} {} 30Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r4ldm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000690000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nmstate-operator-5b5b58f5c8-82r92_openshift-nmstate(7e7d72c9-18b4-4716-aabb-2c55c7070020): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \"/var/tmp/container_images_storage22533152/1\": happened during read: context canceled" logger="UnhandledError" Dec 05 09:20:01 crc kubenswrapper[4815]: E1205 09:20:01.991231 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nmstate-operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: writing blob: storing blob to file \\\"/var/tmp/container_images_storage22533152/1\\\": happened during read: context canceled\"" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92" podUID="7e7d72c9-18b4-4716-aabb-2c55c7070020" Dec 05 09:20:02 crc kubenswrapper[4815]: I1205 09:20:02.558206 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"bda5cae8332591142b5d1ee8c164d1bfba69b735e07e595565abef36f7e9229e"} Dec 05 09:20:02 crc kubenswrapper[4815]: I1205 09:20:02.561589 4815 generic.go:334] "Generic (PLEG): container finished" podID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerID="cff9713891780dd72cf0deae9b26ba0cd1ba06635e6b398f226c83555a65a600" exitCode=0 Dec 05 09:20:02 crc kubenswrapper[4815]: I1205 09:20:02.561645 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w9wt" event={"ID":"732d7c70-b825-4c00-b7d8-d5f6327ebe1e","Type":"ContainerDied","Data":"cff9713891780dd72cf0deae9b26ba0cd1ba06635e6b398f226c83555a65a600"} Dec 05 09:20:02 crc kubenswrapper[4815]: E1205 09:20:02.566937 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nmstate-operator\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/kubernetes-nmstate-rhel9-operator@sha256:dd89e08ed6257597e99b1243839d5c76e6bad72fe9e168c0eba5ce9c449189cf\\\"\"" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92" podUID="7e7d72c9-18b4-4716-aabb-2c55c7070020" Dec 05 09:20:02 crc kubenswrapper[4815]: I1205 09:20:02.890357 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:20:02 crc kubenswrapper[4815]: I1205 09:20:02.931814 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7mhz5\" (UniqueName: \"kubernetes.io/projected/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-kube-api-access-7mhz5\") pod \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " Dec 05 09:20:02 crc kubenswrapper[4815]: I1205 09:20:02.931889 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-catalog-content\") pod \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " Dec 05 09:20:02 crc kubenswrapper[4815]: I1205 09:20:02.931935 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-utilities\") pod \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\" (UID: \"732d7c70-b825-4c00-b7d8-d5f6327ebe1e\") " Dec 05 09:20:02 crc kubenswrapper[4815]: I1205 09:20:02.934046 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-utilities" (OuterVolumeSpecName: "utilities") pod "732d7c70-b825-4c00-b7d8-d5f6327ebe1e" (UID: "732d7c70-b825-4c00-b7d8-d5f6327ebe1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:20:02 crc kubenswrapper[4815]: I1205 09:20:02.940965 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-kube-api-access-7mhz5" (OuterVolumeSpecName: "kube-api-access-7mhz5") pod "732d7c70-b825-4c00-b7d8-d5f6327ebe1e" (UID: "732d7c70-b825-4c00-b7d8-d5f6327ebe1e"). InnerVolumeSpecName "kube-api-access-7mhz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:20:03 crc kubenswrapper[4815]: I1205 09:20:03.033363 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:03 crc kubenswrapper[4815]: I1205 09:20:03.033409 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7mhz5\" (UniqueName: \"kubernetes.io/projected/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-kube-api-access-7mhz5\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:03 crc kubenswrapper[4815]: I1205 09:20:03.037522 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "732d7c70-b825-4c00-b7d8-d5f6327ebe1e" (UID: "732d7c70-b825-4c00-b7d8-d5f6327ebe1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:20:03 crc kubenswrapper[4815]: I1205 09:20:03.135285 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/732d7c70-b825-4c00-b7d8-d5f6327ebe1e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:20:03 crc kubenswrapper[4815]: I1205 09:20:03.570110 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-7w9wt" Dec 05 09:20:03 crc kubenswrapper[4815]: I1205 09:20:03.570138 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-7w9wt" event={"ID":"732d7c70-b825-4c00-b7d8-d5f6327ebe1e","Type":"ContainerDied","Data":"b61d7f61ec758817b835529e9cac4361f3a2f291daf55fb71b1d69754ed5273e"} Dec 05 09:20:03 crc kubenswrapper[4815]: I1205 09:20:03.570776 4815 scope.go:117] "RemoveContainer" containerID="cff9713891780dd72cf0deae9b26ba0cd1ba06635e6b398f226c83555a65a600" Dec 05 09:20:03 crc kubenswrapper[4815]: I1205 09:20:03.599326 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-7w9wt"] Dec 05 09:20:03 crc kubenswrapper[4815]: I1205 09:20:03.599942 4815 scope.go:117] "RemoveContainer" containerID="b3a5d4904edfb5cbd50960e7b33811622311622c3bd74e089e33d43d1f3250c1" Dec 05 09:20:03 crc kubenswrapper[4815]: I1205 09:20:03.612065 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-7w9wt"] Dec 05 09:20:03 crc kubenswrapper[4815]: I1205 09:20:03.635153 4815 scope.go:117] "RemoveContainer" containerID="9731ad28197a6e949ed2eda7641b4faad7859379ea53d8968c1e8bf29969a0d4" Dec 05 09:20:05 crc kubenswrapper[4815]: I1205 09:20:05.429206 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" path="/var/lib/kubelet/pods/732d7c70-b825-4c00-b7d8-d5f6327ebe1e/volumes" Dec 05 09:20:20 crc kubenswrapper[4815]: I1205 09:20:20.697031 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92" event={"ID":"7e7d72c9-18b4-4716-aabb-2c55c7070020","Type":"ContainerStarted","Data":"645f32089b9e677eb63721e4283c8ef609390899ab7eb5d3b8998b70c2280f7f"} Dec 05 09:20:20 crc kubenswrapper[4815]: I1205 09:20:20.717613 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-82r92" podStartSLOduration=1.543791874 podStartE2EDuration="34.717590796s" podCreationTimestamp="2025-12-05 09:19:46 +0000 UTC" firstStartedPulling="2025-12-05 09:19:46.794175743 +0000 UTC m=+785.672782580" lastFinishedPulling="2025-12-05 09:20:19.967974655 +0000 UTC m=+818.846581502" observedRunningTime="2025-12-05 09:20:20.711329496 +0000 UTC m=+819.589936353" watchObservedRunningTime="2025-12-05 09:20:20.717590796 +0000 UTC m=+819.596197633" Dec 05 09:20:21 crc kubenswrapper[4815]: I1205 09:20:21.980709 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq"] Dec 05 09:20:21 crc kubenswrapper[4815]: E1205 09:20:21.980933 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerName="extract-utilities" Dec 05 09:20:21 crc kubenswrapper[4815]: I1205 09:20:21.980945 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerName="extract-utilities" Dec 05 09:20:21 crc kubenswrapper[4815]: E1205 09:20:21.980958 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerName="registry-server" Dec 05 09:20:21 crc kubenswrapper[4815]: I1205 09:20:21.980964 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerName="registry-server" Dec 05 09:20:21 crc kubenswrapper[4815]: E1205 09:20:21.980982 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerName="extract-content" Dec 05 09:20:21 crc kubenswrapper[4815]: I1205 09:20:21.980988 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerName="extract-content" Dec 05 09:20:21 crc kubenswrapper[4815]: I1205 09:20:21.981105 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="732d7c70-b825-4c00-b7d8-d5f6327ebe1e" containerName="registry-server" Dec 05 09:20:21 crc kubenswrapper[4815]: I1205 09:20:21.981924 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq" Dec 05 09:20:21 crc kubenswrapper[4815]: I1205 09:20:21.986786 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-vs8n4" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:21.999999 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq"] Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.001231 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.004249 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.015783 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stcbl\" (UniqueName: \"kubernetes.io/projected/d47f3938-b642-4bd2-b807-b2c962918651-kube-api-access-stcbl\") pod \"nmstate-metrics-7f946cbc9-9frbq\" (UID: \"d47f3938-b642-4bd2-b807-b2c962918651\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.044134 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-d8hkx"] Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.045057 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.075073 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq"] Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.112236 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq"] Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.117833 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stcbl\" (UniqueName: \"kubernetes.io/projected/d47f3938-b642-4bd2-b807-b2c962918651-kube-api-access-stcbl\") pod \"nmstate-metrics-7f946cbc9-9frbq\" (UID: \"d47f3938-b642-4bd2-b807-b2c962918651\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.117891 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-ovs-socket\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.117927 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt2mc\" (UniqueName: \"kubernetes.io/projected/8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5-kube-api-access-vt2mc\") pod \"nmstate-webhook-5f6d4c5ccb-qtprq\" (UID: \"8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.117957 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-nmstate-lock\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.118002 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7vxk\" (UniqueName: \"kubernetes.io/projected/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-kube-api-access-q7vxk\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.118042 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-qtprq\" (UID: \"8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.118128 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-dbus-socket\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.151418 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stcbl\" (UniqueName: \"kubernetes.io/projected/d47f3938-b642-4bd2-b807-b2c962918651-kube-api-access-stcbl\") pod \"nmstate-metrics-7f946cbc9-9frbq\" (UID: \"d47f3938-b642-4bd2-b807-b2c962918651\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.163072 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b"] Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.163723 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.165409 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-r6bxp" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.165596 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.169915 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.174952 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b"] Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.219943 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7vxk\" (UniqueName: \"kubernetes.io/projected/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-kube-api-access-q7vxk\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.220023 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-qtprq\" (UID: \"8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.220059 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-dbus-socket\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.220095 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/028ef3e8-a0e1-443e-840d-0f7244a0e953-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5g45b\" (UID: \"028ef3e8-a0e1-443e-840d-0f7244a0e953\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.220118 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/028ef3e8-a0e1-443e-840d-0f7244a0e953-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5g45b\" (UID: \"028ef3e8-a0e1-443e-840d-0f7244a0e953\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.220159 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26g6n\" (UniqueName: \"kubernetes.io/projected/028ef3e8-a0e1-443e-840d-0f7244a0e953-kube-api-access-26g6n\") pod \"nmstate-console-plugin-7fbb5f6569-5g45b\" (UID: \"028ef3e8-a0e1-443e-840d-0f7244a0e953\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.220823 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-ovs-socket\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.220889 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt2mc\" (UniqueName: \"kubernetes.io/projected/8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5-kube-api-access-vt2mc\") pod \"nmstate-webhook-5f6d4c5ccb-qtprq\" (UID: \"8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.220909 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-nmstate-lock\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.220902 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-dbus-socket\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.220995 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-nmstate-lock\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: E1205 09:20:22.221012 4815 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.221020 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-ovs-socket\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: E1205 09:20:22.221078 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5-tls-key-pair podName:8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5 nodeName:}" failed. No retries permitted until 2025-12-05 09:20:22.721059865 +0000 UTC m=+821.599666702 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-qtprq" (UID: "8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5") : secret "openshift-nmstate-webhook" not found Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.238008 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7vxk\" (UniqueName: \"kubernetes.io/projected/da8c41e3-2e9f-4917-a9e7-efa4debc22c7-kube-api-access-q7vxk\") pod \"nmstate-handler-d8hkx\" (UID: \"da8c41e3-2e9f-4917-a9e7-efa4debc22c7\") " pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.239285 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt2mc\" (UniqueName: \"kubernetes.io/projected/8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5-kube-api-access-vt2mc\") pod \"nmstate-webhook-5f6d4c5ccb-qtprq\" (UID: \"8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.296852 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.321958 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26g6n\" (UniqueName: \"kubernetes.io/projected/028ef3e8-a0e1-443e-840d-0f7244a0e953-kube-api-access-26g6n\") pod \"nmstate-console-plugin-7fbb5f6569-5g45b\" (UID: \"028ef3e8-a0e1-443e-840d-0f7244a0e953\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.322125 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/028ef3e8-a0e1-443e-840d-0f7244a0e953-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5g45b\" (UID: \"028ef3e8-a0e1-443e-840d-0f7244a0e953\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.322162 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/028ef3e8-a0e1-443e-840d-0f7244a0e953-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5g45b\" (UID: \"028ef3e8-a0e1-443e-840d-0f7244a0e953\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:22 crc kubenswrapper[4815]: E1205 09:20:22.322356 4815 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 05 09:20:22 crc kubenswrapper[4815]: E1205 09:20:22.322440 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/028ef3e8-a0e1-443e-840d-0f7244a0e953-plugin-serving-cert podName:028ef3e8-a0e1-443e-840d-0f7244a0e953 nodeName:}" failed. No retries permitted until 2025-12-05 09:20:22.822401406 +0000 UTC m=+821.701008253 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/028ef3e8-a0e1-443e-840d-0f7244a0e953-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-5g45b" (UID: "028ef3e8-a0e1-443e-840d-0f7244a0e953") : secret "plugin-serving-cert" not found Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.323722 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/028ef3e8-a0e1-443e-840d-0f7244a0e953-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-5g45b\" (UID: \"028ef3e8-a0e1-443e-840d-0f7244a0e953\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.347485 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26g6n\" (UniqueName: \"kubernetes.io/projected/028ef3e8-a0e1-443e-840d-0f7244a0e953-kube-api-access-26g6n\") pod \"nmstate-console-plugin-7fbb5f6569-5g45b\" (UID: \"028ef3e8-a0e1-443e-840d-0f7244a0e953\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.369251 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:22 crc kubenswrapper[4815]: W1205 09:20:22.419121 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda8c41e3_2e9f_4917_a9e7_efa4debc22c7.slice/crio-c7312a67e24c221fe1a0fbd366262d2e980bcc9fc6bc47ccecef2d9e9ec6b6bc WatchSource:0}: Error finding container c7312a67e24c221fe1a0fbd366262d2e980bcc9fc6bc47ccecef2d9e9ec6b6bc: Status 404 returned error can't find the container with id c7312a67e24c221fe1a0fbd366262d2e980bcc9fc6bc47ccecef2d9e9ec6b6bc Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.452611 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-796f78c94d-pklzr"] Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.457820 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.459289 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-796f78c94d-pklzr"] Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.525809 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-console-config\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.525873 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/66ec25b4-230b-412e-960a-02072a54290f-console-oauth-config\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.525910 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-service-ca\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.525929 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8jgt\" (UniqueName: \"kubernetes.io/projected/66ec25b4-230b-412e-960a-02072a54290f-kube-api-access-s8jgt\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.525962 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-trusted-ca-bundle\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.525988 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-oauth-serving-cert\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.526002 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/66ec25b4-230b-412e-960a-02072a54290f-console-serving-cert\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.627379 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/66ec25b4-230b-412e-960a-02072a54290f-console-oauth-config\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.627665 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-service-ca\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.627685 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8jgt\" (UniqueName: \"kubernetes.io/projected/66ec25b4-230b-412e-960a-02072a54290f-kube-api-access-s8jgt\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.627730 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-trusted-ca-bundle\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.627759 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/66ec25b4-230b-412e-960a-02072a54290f-console-serving-cert\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.627777 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-oauth-serving-cert\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.627805 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-console-config\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.628648 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-service-ca\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.628736 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-console-config\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.628973 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-oauth-serving-cert\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.629194 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/66ec25b4-230b-412e-960a-02072a54290f-trusted-ca-bundle\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.632577 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/66ec25b4-230b-412e-960a-02072a54290f-console-oauth-config\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.634345 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/66ec25b4-230b-412e-960a-02072a54290f-console-serving-cert\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.635964 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq"] Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.646879 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8jgt\" (UniqueName: \"kubernetes.io/projected/66ec25b4-230b-412e-960a-02072a54290f-kube-api-access-s8jgt\") pod \"console-796f78c94d-pklzr\" (UID: \"66ec25b4-230b-412e-960a-02072a54290f\") " pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.706033 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq" event={"ID":"d47f3938-b642-4bd2-b807-b2c962918651","Type":"ContainerStarted","Data":"25b6d75b94df8e17efa76b2de76e5db20c81efdda3d02d12ea173b66f8c2fb52"} Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.707005 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-d8hkx" event={"ID":"da8c41e3-2e9f-4917-a9e7-efa4debc22c7","Type":"ContainerStarted","Data":"c7312a67e24c221fe1a0fbd366262d2e980bcc9fc6bc47ccecef2d9e9ec6b6bc"} Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.729257 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-qtprq\" (UID: \"8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.732375 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-qtprq\" (UID: \"8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.830807 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/028ef3e8-a0e1-443e-840d-0f7244a0e953-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5g45b\" (UID: \"028ef3e8-a0e1-443e-840d-0f7244a0e953\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.833724 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/028ef3e8-a0e1-443e-840d-0f7244a0e953-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-5g45b\" (UID: \"028ef3e8-a0e1-443e-840d-0f7244a0e953\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.845758 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:22 crc kubenswrapper[4815]: I1205 09:20:22.933761 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" Dec 05 09:20:23 crc kubenswrapper[4815]: I1205 09:20:23.052795 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-796f78c94d-pklzr"] Dec 05 09:20:23 crc kubenswrapper[4815]: W1205 09:20:23.062834 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod66ec25b4_230b_412e_960a_02072a54290f.slice/crio-21caaf870886a512b7352512bb4ec0766b2868372cedd812fb2c84a58fb82cd1 WatchSource:0}: Error finding container 21caaf870886a512b7352512bb4ec0766b2868372cedd812fb2c84a58fb82cd1: Status 404 returned error can't find the container with id 21caaf870886a512b7352512bb4ec0766b2868372cedd812fb2c84a58fb82cd1 Dec 05 09:20:23 crc kubenswrapper[4815]: I1205 09:20:23.090624 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" Dec 05 09:20:23 crc kubenswrapper[4815]: I1205 09:20:23.139244 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq"] Dec 05 09:20:23 crc kubenswrapper[4815]: I1205 09:20:23.290471 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b"] Dec 05 09:20:23 crc kubenswrapper[4815]: I1205 09:20:23.718144 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-796f78c94d-pklzr" event={"ID":"66ec25b4-230b-412e-960a-02072a54290f","Type":"ContainerStarted","Data":"fb49ab26bfc241e454240ca565ba75261da42c204d2029f176213ebc9de0311a"} Dec 05 09:20:23 crc kubenswrapper[4815]: I1205 09:20:23.718198 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-796f78c94d-pklzr" event={"ID":"66ec25b4-230b-412e-960a-02072a54290f","Type":"ContainerStarted","Data":"21caaf870886a512b7352512bb4ec0766b2868372cedd812fb2c84a58fb82cd1"} Dec 05 09:20:23 crc kubenswrapper[4815]: I1205 09:20:23.721293 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" event={"ID":"028ef3e8-a0e1-443e-840d-0f7244a0e953","Type":"ContainerStarted","Data":"14a9b9f4bfae57c0339c877a7ce4b7308b5cffba5d6b17d29ea2f22f18b4363a"} Dec 05 09:20:23 crc kubenswrapper[4815]: I1205 09:20:23.724227 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" event={"ID":"8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5","Type":"ContainerStarted","Data":"419c291cfac59c8e595426657c6d05df308c408b91ad05467a208dd98229e992"} Dec 05 09:20:23 crc kubenswrapper[4815]: I1205 09:20:23.739964 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-796f78c94d-pklzr" podStartSLOduration=1.739942764 podStartE2EDuration="1.739942764s" podCreationTimestamp="2025-12-05 09:20:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:20:23.734815394 +0000 UTC m=+822.613422251" watchObservedRunningTime="2025-12-05 09:20:23.739942764 +0000 UTC m=+822.618549611" Dec 05 09:20:26 crc kubenswrapper[4815]: I1205 09:20:26.741677 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" event={"ID":"8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5","Type":"ContainerStarted","Data":"7f1df1c35f5a2d42dae79073a271cbf9f4e7005c4972bdef78e323e2cba9400c"} Dec 05 09:20:26 crc kubenswrapper[4815]: I1205 09:20:26.742203 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" Dec 05 09:20:26 crc kubenswrapper[4815]: I1205 09:20:26.743203 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq" event={"ID":"d47f3938-b642-4bd2-b807-b2c962918651","Type":"ContainerStarted","Data":"adc6f11633bbf15e9a497cc0d2a2bd1d9fbf08516e4e71fae67430fcc1d6b762"} Dec 05 09:20:26 crc kubenswrapper[4815]: I1205 09:20:26.744462 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-d8hkx" event={"ID":"da8c41e3-2e9f-4917-a9e7-efa4debc22c7","Type":"ContainerStarted","Data":"ae0db5cc10e493168b634800a70fd39d6369e5cbc30a1e9d66384cd40067de5c"} Dec 05 09:20:26 crc kubenswrapper[4815]: I1205 09:20:26.744568 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:26 crc kubenswrapper[4815]: I1205 09:20:26.746186 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" event={"ID":"028ef3e8-a0e1-443e-840d-0f7244a0e953","Type":"ContainerStarted","Data":"ddcff8f4ab13fa38ec26e9b5f8df8062ccf095a0f0b5a6aa004ba5f670e93a22"} Dec 05 09:20:26 crc kubenswrapper[4815]: I1205 09:20:26.762808 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" podStartSLOduration=3.157958591 podStartE2EDuration="5.762789175s" podCreationTimestamp="2025-12-05 09:20:21 +0000 UTC" firstStartedPulling="2025-12-05 09:20:23.175758344 +0000 UTC m=+822.054365181" lastFinishedPulling="2025-12-05 09:20:25.780588928 +0000 UTC m=+824.659195765" observedRunningTime="2025-12-05 09:20:26.760117113 +0000 UTC m=+825.638723960" watchObservedRunningTime="2025-12-05 09:20:26.762789175 +0000 UTC m=+825.641396012" Dec 05 09:20:26 crc kubenswrapper[4815]: I1205 09:20:26.795734 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-d8hkx" podStartSLOduration=0.940459264 podStartE2EDuration="4.795715443s" podCreationTimestamp="2025-12-05 09:20:22 +0000 UTC" firstStartedPulling="2025-12-05 09:20:22.421539077 +0000 UTC m=+821.300145914" lastFinishedPulling="2025-12-05 09:20:26.276795256 +0000 UTC m=+825.155402093" observedRunningTime="2025-12-05 09:20:26.791406546 +0000 UTC m=+825.670013403" watchObservedRunningTime="2025-12-05 09:20:26.795715443 +0000 UTC m=+825.674322280" Dec 05 09:20:26 crc kubenswrapper[4815]: I1205 09:20:26.813454 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-5g45b" podStartSLOduration=1.724954136 podStartE2EDuration="4.813436785s" podCreationTimestamp="2025-12-05 09:20:22 +0000 UTC" firstStartedPulling="2025-12-05 09:20:23.293352288 +0000 UTC m=+822.171959125" lastFinishedPulling="2025-12-05 09:20:26.381834937 +0000 UTC m=+825.260441774" observedRunningTime="2025-12-05 09:20:26.812219452 +0000 UTC m=+825.690826299" watchObservedRunningTime="2025-12-05 09:20:26.813436785 +0000 UTC m=+825.692043622" Dec 05 09:20:32 crc kubenswrapper[4815]: I1205 09:20:32.399912 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-d8hkx" Dec 05 09:20:32 crc kubenswrapper[4815]: I1205 09:20:32.846199 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:32 crc kubenswrapper[4815]: I1205 09:20:32.846260 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:32 crc kubenswrapper[4815]: I1205 09:20:32.851187 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:33 crc kubenswrapper[4815]: I1205 09:20:33.801870 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq" event={"ID":"d47f3938-b642-4bd2-b807-b2c962918651","Type":"ContainerStarted","Data":"b12dc6e5306276c08d67c61033fab5d595d10c4c59fbfe8a79ed8e44fe9ab0f5"} Dec 05 09:20:33 crc kubenswrapper[4815]: I1205 09:20:33.810243 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-796f78c94d-pklzr" Dec 05 09:20:33 crc kubenswrapper[4815]: I1205 09:20:33.824650 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-9frbq" podStartSLOduration=2.060993157 podStartE2EDuration="12.824631303s" podCreationTimestamp="2025-12-05 09:20:21 +0000 UTC" firstStartedPulling="2025-12-05 09:20:22.646971608 +0000 UTC m=+821.525578445" lastFinishedPulling="2025-12-05 09:20:33.410609734 +0000 UTC m=+832.289216591" observedRunningTime="2025-12-05 09:20:33.820731227 +0000 UTC m=+832.699338064" watchObservedRunningTime="2025-12-05 09:20:33.824631303 +0000 UTC m=+832.703238150" Dec 05 09:20:33 crc kubenswrapper[4815]: I1205 09:20:33.886154 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-d2j7q"] Dec 05 09:20:42 crc kubenswrapper[4815]: I1205 09:20:42.939986 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-qtprq" Dec 05 09:20:58 crc kubenswrapper[4815]: I1205 09:20:58.933939 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-d2j7q" podUID="bf95c26c-10fd-41bb-968a-2b5df4066d29" containerName="console" containerID="cri-o://da63efd2ce136aaafcc6a510282ceefd944b7bbc24ee5b12a81ab2e5a4d203fa" gracePeriod=15 Dec 05 09:20:59 crc kubenswrapper[4815]: I1205 09:20:59.991404 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-d2j7q_bf95c26c-10fd-41bb-968a-2b5df4066d29/console/0.log" Dec 05 09:20:59 crc kubenswrapper[4815]: I1205 09:20:59.991761 4815 generic.go:334] "Generic (PLEG): container finished" podID="bf95c26c-10fd-41bb-968a-2b5df4066d29" containerID="da63efd2ce136aaafcc6a510282ceefd944b7bbc24ee5b12a81ab2e5a4d203fa" exitCode=2 Dec 05 09:20:59 crc kubenswrapper[4815]: I1205 09:20:59.991808 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-d2j7q" event={"ID":"bf95c26c-10fd-41bb-968a-2b5df4066d29","Type":"ContainerDied","Data":"da63efd2ce136aaafcc6a510282ceefd944b7bbc24ee5b12a81ab2e5a4d203fa"} Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.528032 4815 patch_prober.go:28] interesting pod/console-f9d7485db-d2j7q container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.528109 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-f9d7485db-d2j7q" podUID="bf95c26c-10fd-41bb-968a-2b5df4066d29" containerName="console" probeResult="failure" output="Get \"https://10.217.0.9:8443/health\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.736285 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-d2j7q_bf95c26c-10fd-41bb-968a-2b5df4066d29/console/0.log" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.736568 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.783257 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp8tg\" (UniqueName: \"kubernetes.io/projected/bf95c26c-10fd-41bb-968a-2b5df4066d29-kube-api-access-cp8tg\") pod \"bf95c26c-10fd-41bb-968a-2b5df4066d29\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.783311 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-service-ca\") pod \"bf95c26c-10fd-41bb-968a-2b5df4066d29\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.783354 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-serving-cert\") pod \"bf95c26c-10fd-41bb-968a-2b5df4066d29\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.783389 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-oauth-serving-cert\") pod \"bf95c26c-10fd-41bb-968a-2b5df4066d29\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.783429 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-config\") pod \"bf95c26c-10fd-41bb-968a-2b5df4066d29\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.785211 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "bf95c26c-10fd-41bb-968a-2b5df4066d29" (UID: "bf95c26c-10fd-41bb-968a-2b5df4066d29"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.785236 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-service-ca" (OuterVolumeSpecName: "service-ca") pod "bf95c26c-10fd-41bb-968a-2b5df4066d29" (UID: "bf95c26c-10fd-41bb-968a-2b5df4066d29"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.783459 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-oauth-config\") pod \"bf95c26c-10fd-41bb-968a-2b5df4066d29\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.785425 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-trusted-ca-bundle\") pod \"bf95c26c-10fd-41bb-968a-2b5df4066d29\" (UID: \"bf95c26c-10fd-41bb-968a-2b5df4066d29\") " Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.785766 4815 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-service-ca\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.785781 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-config" (OuterVolumeSpecName: "console-config") pod "bf95c26c-10fd-41bb-968a-2b5df4066d29" (UID: "bf95c26c-10fd-41bb-968a-2b5df4066d29"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.785793 4815 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.786320 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "bf95c26c-10fd-41bb-968a-2b5df4066d29" (UID: "bf95c26c-10fd-41bb-968a-2b5df4066d29"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.800456 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf95c26c-10fd-41bb-968a-2b5df4066d29-kube-api-access-cp8tg" (OuterVolumeSpecName: "kube-api-access-cp8tg") pod "bf95c26c-10fd-41bb-968a-2b5df4066d29" (UID: "bf95c26c-10fd-41bb-968a-2b5df4066d29"). InnerVolumeSpecName "kube-api-access-cp8tg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.800481 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "bf95c26c-10fd-41bb-968a-2b5df4066d29" (UID: "bf95c26c-10fd-41bb-968a-2b5df4066d29"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.800976 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "bf95c26c-10fd-41bb-968a-2b5df4066d29" (UID: "bf95c26c-10fd-41bb-968a-2b5df4066d29"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.886533 4815 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.886570 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp8tg\" (UniqueName: \"kubernetes.io/projected/bf95c26c-10fd-41bb-968a-2b5df4066d29-kube-api-access-cp8tg\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.886581 4815 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.886590 4815 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.886599 4815 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/bf95c26c-10fd-41bb-968a-2b5df4066d29-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:00 crc kubenswrapper[4815]: I1205 09:21:00.999689 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-d2j7q_bf95c26c-10fd-41bb-968a-2b5df4066d29/console/0.log" Dec 05 09:21:01 crc kubenswrapper[4815]: I1205 09:21:00.999741 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-d2j7q" event={"ID":"bf95c26c-10fd-41bb-968a-2b5df4066d29","Type":"ContainerDied","Data":"de0a88d4076e535b7684a5eae61247dcdf9bf3ca04c95d978beca6abd51cbb43"} Dec 05 09:21:01 crc kubenswrapper[4815]: I1205 09:21:00.999774 4815 scope.go:117] "RemoveContainer" containerID="da63efd2ce136aaafcc6a510282ceefd944b7bbc24ee5b12a81ab2e5a4d203fa" Dec 05 09:21:01 crc kubenswrapper[4815]: I1205 09:21:00.999836 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-d2j7q" Dec 05 09:21:01 crc kubenswrapper[4815]: I1205 09:21:01.034013 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-d2j7q"] Dec 05 09:21:01 crc kubenswrapper[4815]: I1205 09:21:01.039105 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-d2j7q"] Dec 05 09:21:01 crc kubenswrapper[4815]: I1205 09:21:01.428619 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf95c26c-10fd-41bb-968a-2b5df4066d29" path="/var/lib/kubelet/pods/bf95c26c-10fd-41bb-968a-2b5df4066d29/volumes" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.276065 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw"] Dec 05 09:21:07 crc kubenswrapper[4815]: E1205 09:21:07.277165 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf95c26c-10fd-41bb-968a-2b5df4066d29" containerName="console" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.277181 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf95c26c-10fd-41bb-968a-2b5df4066d29" containerName="console" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.277310 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf95c26c-10fd-41bb-968a-2b5df4066d29" containerName="console" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.278300 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.281021 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.285663 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw"] Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.369400 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.369725 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29cxq\" (UniqueName: \"kubernetes.io/projected/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-kube-api-access-29cxq\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.369833 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.470642 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29cxq\" (UniqueName: \"kubernetes.io/projected/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-kube-api-access-29cxq\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.470692 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.470727 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.471378 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.471665 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.501915 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29cxq\" (UniqueName: \"kubernetes.io/projected/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-kube-api-access-29cxq\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.592740 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:07 crc kubenswrapper[4815]: I1205 09:21:07.800067 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw"] Dec 05 09:21:08 crc kubenswrapper[4815]: I1205 09:21:08.042902 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" event={"ID":"1f7f5e67-9cbf-4059-9afd-ea5224d6e571","Type":"ContainerStarted","Data":"27067f18047118aa786ec7a1763720c5391d894a28223823af628dc151d4adb9"} Dec 05 09:21:10 crc kubenswrapper[4815]: I1205 09:21:10.057705 4815 generic.go:334] "Generic (PLEG): container finished" podID="1f7f5e67-9cbf-4059-9afd-ea5224d6e571" containerID="682c630155af63a1ec523af109303ecdc22a8e14a5167f53d4389a4c95157883" exitCode=0 Dec 05 09:21:10 crc kubenswrapper[4815]: I1205 09:21:10.057749 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" event={"ID":"1f7f5e67-9cbf-4059-9afd-ea5224d6e571","Type":"ContainerDied","Data":"682c630155af63a1ec523af109303ecdc22a8e14a5167f53d4389a4c95157883"} Dec 05 09:21:13 crc kubenswrapper[4815]: I1205 09:21:13.077568 4815 generic.go:334] "Generic (PLEG): container finished" podID="1f7f5e67-9cbf-4059-9afd-ea5224d6e571" containerID="08fa27414e0ced3f029ebc33595ae6c7bd2051b9759f42de17583bf92c5beb36" exitCode=0 Dec 05 09:21:13 crc kubenswrapper[4815]: I1205 09:21:13.077668 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" event={"ID":"1f7f5e67-9cbf-4059-9afd-ea5224d6e571","Type":"ContainerDied","Data":"08fa27414e0ced3f029ebc33595ae6c7bd2051b9759f42de17583bf92c5beb36"} Dec 05 09:21:14 crc kubenswrapper[4815]: I1205 09:21:14.087321 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" event={"ID":"1f7f5e67-9cbf-4059-9afd-ea5224d6e571","Type":"ContainerStarted","Data":"3ad53e7d86965384a4b60f5e8bb12dd6339fd1cf983fe14f4b2ffde7e5d0811f"} Dec 05 09:21:14 crc kubenswrapper[4815]: I1205 09:21:14.105644 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" podStartSLOduration=4.590875122 podStartE2EDuration="7.105626644s" podCreationTimestamp="2025-12-05 09:21:07 +0000 UTC" firstStartedPulling="2025-12-05 09:21:10.059201606 +0000 UTC m=+868.937808443" lastFinishedPulling="2025-12-05 09:21:12.573953128 +0000 UTC m=+871.452559965" observedRunningTime="2025-12-05 09:21:14.104155424 +0000 UTC m=+872.982762261" watchObservedRunningTime="2025-12-05 09:21:14.105626644 +0000 UTC m=+872.984233481" Dec 05 09:21:15 crc kubenswrapper[4815]: I1205 09:21:15.095316 4815 generic.go:334] "Generic (PLEG): container finished" podID="1f7f5e67-9cbf-4059-9afd-ea5224d6e571" containerID="3ad53e7d86965384a4b60f5e8bb12dd6339fd1cf983fe14f4b2ffde7e5d0811f" exitCode=0 Dec 05 09:21:15 crc kubenswrapper[4815]: I1205 09:21:15.095359 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" event={"ID":"1f7f5e67-9cbf-4059-9afd-ea5224d6e571","Type":"ContainerDied","Data":"3ad53e7d86965384a4b60f5e8bb12dd6339fd1cf983fe14f4b2ffde7e5d0811f"} Dec 05 09:21:16 crc kubenswrapper[4815]: I1205 09:21:16.354579 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:16 crc kubenswrapper[4815]: I1205 09:21:16.508222 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-bundle\") pod \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " Dec 05 09:21:16 crc kubenswrapper[4815]: I1205 09:21:16.508333 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-util\") pod \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " Dec 05 09:21:16 crc kubenswrapper[4815]: I1205 09:21:16.508424 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29cxq\" (UniqueName: \"kubernetes.io/projected/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-kube-api-access-29cxq\") pod \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\" (UID: \"1f7f5e67-9cbf-4059-9afd-ea5224d6e571\") " Dec 05 09:21:16 crc kubenswrapper[4815]: I1205 09:21:16.509823 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-bundle" (OuterVolumeSpecName: "bundle") pod "1f7f5e67-9cbf-4059-9afd-ea5224d6e571" (UID: "1f7f5e67-9cbf-4059-9afd-ea5224d6e571"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:21:16 crc kubenswrapper[4815]: I1205 09:21:16.516848 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-kube-api-access-29cxq" (OuterVolumeSpecName: "kube-api-access-29cxq") pod "1f7f5e67-9cbf-4059-9afd-ea5224d6e571" (UID: "1f7f5e67-9cbf-4059-9afd-ea5224d6e571"). InnerVolumeSpecName "kube-api-access-29cxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:21:16 crc kubenswrapper[4815]: I1205 09:21:16.519706 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-util" (OuterVolumeSpecName: "util") pod "1f7f5e67-9cbf-4059-9afd-ea5224d6e571" (UID: "1f7f5e67-9cbf-4059-9afd-ea5224d6e571"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:21:16 crc kubenswrapper[4815]: I1205 09:21:16.609713 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29cxq\" (UniqueName: \"kubernetes.io/projected/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-kube-api-access-29cxq\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:16 crc kubenswrapper[4815]: I1205 09:21:16.609746 4815 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:16 crc kubenswrapper[4815]: I1205 09:21:16.609757 4815 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1f7f5e67-9cbf-4059-9afd-ea5224d6e571-util\") on node \"crc\" DevicePath \"\"" Dec 05 09:21:17 crc kubenswrapper[4815]: I1205 09:21:17.108518 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" event={"ID":"1f7f5e67-9cbf-4059-9afd-ea5224d6e571","Type":"ContainerDied","Data":"27067f18047118aa786ec7a1763720c5391d894a28223823af628dc151d4adb9"} Dec 05 09:21:17 crc kubenswrapper[4815]: I1205 09:21:17.108858 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27067f18047118aa786ec7a1763720c5391d894a28223823af628dc151d4adb9" Dec 05 09:21:17 crc kubenswrapper[4815]: I1205 09:21:17.108581 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.605062 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74"] Dec 05 09:21:31 crc kubenswrapper[4815]: E1205 09:21:31.605752 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f7f5e67-9cbf-4059-9afd-ea5224d6e571" containerName="extract" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.605765 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f7f5e67-9cbf-4059-9afd-ea5224d6e571" containerName="extract" Dec 05 09:21:31 crc kubenswrapper[4815]: E1205 09:21:31.605775 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f7f5e67-9cbf-4059-9afd-ea5224d6e571" containerName="pull" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.605781 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f7f5e67-9cbf-4059-9afd-ea5224d6e571" containerName="pull" Dec 05 09:21:31 crc kubenswrapper[4815]: E1205 09:21:31.605797 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f7f5e67-9cbf-4059-9afd-ea5224d6e571" containerName="util" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.605803 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f7f5e67-9cbf-4059-9afd-ea5224d6e571" containerName="util" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.605925 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f7f5e67-9cbf-4059-9afd-ea5224d6e571" containerName="extract" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.606299 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.610454 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.610740 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-6s9zq" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.610892 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.611138 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.617356 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.644364 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74"] Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.659034 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37e3bb94-27f2-4bfe-9c26-4bbb3025330c-webhook-cert\") pod \"metallb-operator-controller-manager-7f5859557b-zxp74\" (UID: \"37e3bb94-27f2-4bfe-9c26-4bbb3025330c\") " pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.659086 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37e3bb94-27f2-4bfe-9c26-4bbb3025330c-apiservice-cert\") pod \"metallb-operator-controller-manager-7f5859557b-zxp74\" (UID: \"37e3bb94-27f2-4bfe-9c26-4bbb3025330c\") " pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.659131 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pv6c\" (UniqueName: \"kubernetes.io/projected/37e3bb94-27f2-4bfe-9c26-4bbb3025330c-kube-api-access-8pv6c\") pod \"metallb-operator-controller-manager-7f5859557b-zxp74\" (UID: \"37e3bb94-27f2-4bfe-9c26-4bbb3025330c\") " pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.760400 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37e3bb94-27f2-4bfe-9c26-4bbb3025330c-apiservice-cert\") pod \"metallb-operator-controller-manager-7f5859557b-zxp74\" (UID: \"37e3bb94-27f2-4bfe-9c26-4bbb3025330c\") " pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.760706 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pv6c\" (UniqueName: \"kubernetes.io/projected/37e3bb94-27f2-4bfe-9c26-4bbb3025330c-kube-api-access-8pv6c\") pod \"metallb-operator-controller-manager-7f5859557b-zxp74\" (UID: \"37e3bb94-27f2-4bfe-9c26-4bbb3025330c\") " pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.760858 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37e3bb94-27f2-4bfe-9c26-4bbb3025330c-webhook-cert\") pod \"metallb-operator-controller-manager-7f5859557b-zxp74\" (UID: \"37e3bb94-27f2-4bfe-9c26-4bbb3025330c\") " pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.766847 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/37e3bb94-27f2-4bfe-9c26-4bbb3025330c-webhook-cert\") pod \"metallb-operator-controller-manager-7f5859557b-zxp74\" (UID: \"37e3bb94-27f2-4bfe-9c26-4bbb3025330c\") " pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.767502 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/37e3bb94-27f2-4bfe-9c26-4bbb3025330c-apiservice-cert\") pod \"metallb-operator-controller-manager-7f5859557b-zxp74\" (UID: \"37e3bb94-27f2-4bfe-9c26-4bbb3025330c\") " pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.791789 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pv6c\" (UniqueName: \"kubernetes.io/projected/37e3bb94-27f2-4bfe-9c26-4bbb3025330c-kube-api-access-8pv6c\") pod \"metallb-operator-controller-manager-7f5859557b-zxp74\" (UID: \"37e3bb94-27f2-4bfe-9c26-4bbb3025330c\") " pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.916453 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-69858568cc-scmsp"] Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.917358 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.920118 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.920409 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.921321 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-pkvls" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.922574 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.933416 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-69858568cc-scmsp"] Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.966672 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1aabbb7d-8736-4ff3-a41e-7599a4e74194-webhook-cert\") pod \"metallb-operator-webhook-server-69858568cc-scmsp\" (UID: \"1aabbb7d-8736-4ff3-a41e-7599a4e74194\") " pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.966773 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkqgz\" (UniqueName: \"kubernetes.io/projected/1aabbb7d-8736-4ff3-a41e-7599a4e74194-kube-api-access-rkqgz\") pod \"metallb-operator-webhook-server-69858568cc-scmsp\" (UID: \"1aabbb7d-8736-4ff3-a41e-7599a4e74194\") " pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:31 crc kubenswrapper[4815]: I1205 09:21:31.966873 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1aabbb7d-8736-4ff3-a41e-7599a4e74194-apiservice-cert\") pod \"metallb-operator-webhook-server-69858568cc-scmsp\" (UID: \"1aabbb7d-8736-4ff3-a41e-7599a4e74194\") " pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:32 crc kubenswrapper[4815]: I1205 09:21:32.068274 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkqgz\" (UniqueName: \"kubernetes.io/projected/1aabbb7d-8736-4ff3-a41e-7599a4e74194-kube-api-access-rkqgz\") pod \"metallb-operator-webhook-server-69858568cc-scmsp\" (UID: \"1aabbb7d-8736-4ff3-a41e-7599a4e74194\") " pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:32 crc kubenswrapper[4815]: I1205 09:21:32.068362 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1aabbb7d-8736-4ff3-a41e-7599a4e74194-apiservice-cert\") pod \"metallb-operator-webhook-server-69858568cc-scmsp\" (UID: \"1aabbb7d-8736-4ff3-a41e-7599a4e74194\") " pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:32 crc kubenswrapper[4815]: I1205 09:21:32.068403 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1aabbb7d-8736-4ff3-a41e-7599a4e74194-webhook-cert\") pod \"metallb-operator-webhook-server-69858568cc-scmsp\" (UID: \"1aabbb7d-8736-4ff3-a41e-7599a4e74194\") " pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:32 crc kubenswrapper[4815]: I1205 09:21:32.073827 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1aabbb7d-8736-4ff3-a41e-7599a4e74194-webhook-cert\") pod \"metallb-operator-webhook-server-69858568cc-scmsp\" (UID: \"1aabbb7d-8736-4ff3-a41e-7599a4e74194\") " pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:32 crc kubenswrapper[4815]: I1205 09:21:32.090905 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1aabbb7d-8736-4ff3-a41e-7599a4e74194-apiservice-cert\") pod \"metallb-operator-webhook-server-69858568cc-scmsp\" (UID: \"1aabbb7d-8736-4ff3-a41e-7599a4e74194\") " pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:32 crc kubenswrapper[4815]: I1205 09:21:32.098621 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkqgz\" (UniqueName: \"kubernetes.io/projected/1aabbb7d-8736-4ff3-a41e-7599a4e74194-kube-api-access-rkqgz\") pod \"metallb-operator-webhook-server-69858568cc-scmsp\" (UID: \"1aabbb7d-8736-4ff3-a41e-7599a4e74194\") " pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:32 crc kubenswrapper[4815]: I1205 09:21:32.243302 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:32 crc kubenswrapper[4815]: I1205 09:21:32.509184 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74"] Dec 05 09:21:32 crc kubenswrapper[4815]: W1205 09:21:32.523076 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37e3bb94_27f2_4bfe_9c26_4bbb3025330c.slice/crio-0c34b97f414c5cf77ec503662c76441a2f23de754bb0548f435f4e7648a3b11c WatchSource:0}: Error finding container 0c34b97f414c5cf77ec503662c76441a2f23de754bb0548f435f4e7648a3b11c: Status 404 returned error can't find the container with id 0c34b97f414c5cf77ec503662c76441a2f23de754bb0548f435f4e7648a3b11c Dec 05 09:21:32 crc kubenswrapper[4815]: I1205 09:21:32.681156 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-69858568cc-scmsp"] Dec 05 09:21:33 crc kubenswrapper[4815]: I1205 09:21:33.300103 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" event={"ID":"37e3bb94-27f2-4bfe-9c26-4bbb3025330c","Type":"ContainerStarted","Data":"0c34b97f414c5cf77ec503662c76441a2f23de754bb0548f435f4e7648a3b11c"} Dec 05 09:21:33 crc kubenswrapper[4815]: I1205 09:21:33.301548 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" event={"ID":"1aabbb7d-8736-4ff3-a41e-7599a4e74194","Type":"ContainerStarted","Data":"7c52da3faf4f189aafe069e43ac42d4f3d6aa46db856485c3ca9292f92221a88"} Dec 05 09:21:49 crc kubenswrapper[4815]: I1205 09:21:49.403222 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" event={"ID":"1aabbb7d-8736-4ff3-a41e-7599a4e74194","Type":"ContainerStarted","Data":"a869e6321b4d90e448211f5451aa9f239e3291d5586f27596aa0d8b430aa1206"} Dec 05 09:21:49 crc kubenswrapper[4815]: I1205 09:21:49.404653 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" event={"ID":"37e3bb94-27f2-4bfe-9c26-4bbb3025330c","Type":"ContainerStarted","Data":"00f5a7ab64bb1fb21bf481bd42dcc068b56b94dc5dca0556a6be663f77b5fb20"} Dec 05 09:21:49 crc kubenswrapper[4815]: I1205 09:21:49.404722 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:21:49 crc kubenswrapper[4815]: I1205 09:21:49.404934 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:21:49 crc kubenswrapper[4815]: I1205 09:21:49.427727 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" podStartSLOduration=2.695371234 podStartE2EDuration="18.427706081s" podCreationTimestamp="2025-12-05 09:21:31 +0000 UTC" firstStartedPulling="2025-12-05 09:21:32.695834198 +0000 UTC m=+891.574441025" lastFinishedPulling="2025-12-05 09:21:48.428169035 +0000 UTC m=+907.306775872" observedRunningTime="2025-12-05 09:21:49.423689332 +0000 UTC m=+908.302296169" watchObservedRunningTime="2025-12-05 09:21:49.427706081 +0000 UTC m=+908.306312918" Dec 05 09:21:49 crc kubenswrapper[4815]: I1205 09:21:49.535525 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" podStartSLOduration=2.649758774 podStartE2EDuration="18.535476912s" podCreationTimestamp="2025-12-05 09:21:31 +0000 UTC" firstStartedPulling="2025-12-05 09:21:32.536465793 +0000 UTC m=+891.415072620" lastFinishedPulling="2025-12-05 09:21:48.422183921 +0000 UTC m=+907.300790758" observedRunningTime="2025-12-05 09:21:49.528967594 +0000 UTC m=+908.407574431" watchObservedRunningTime="2025-12-05 09:21:49.535476912 +0000 UTC m=+908.414083749" Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.416970 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4j7bd"] Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.418576 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.504887 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4j7bd"] Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.558417 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-utilities\") pod \"certified-operators-4j7bd\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.559201 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-catalog-content\") pod \"certified-operators-4j7bd\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.559250 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbs5f\" (UniqueName: \"kubernetes.io/projected/de650b0f-13f4-4621-9026-d812aec9a752-kube-api-access-gbs5f\") pod \"certified-operators-4j7bd\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.660199 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-utilities\") pod \"certified-operators-4j7bd\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.660279 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-catalog-content\") pod \"certified-operators-4j7bd\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.660299 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbs5f\" (UniqueName: \"kubernetes.io/projected/de650b0f-13f4-4621-9026-d812aec9a752-kube-api-access-gbs5f\") pod \"certified-operators-4j7bd\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.660969 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-utilities\") pod \"certified-operators-4j7bd\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.661108 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-catalog-content\") pod \"certified-operators-4j7bd\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.700264 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbs5f\" (UniqueName: \"kubernetes.io/projected/de650b0f-13f4-4621-9026-d812aec9a752-kube-api-access-gbs5f\") pod \"certified-operators-4j7bd\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:21:58 crc kubenswrapper[4815]: I1205 09:21:58.780447 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:21:59 crc kubenswrapper[4815]: I1205 09:21:59.160114 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4j7bd"] Dec 05 09:21:59 crc kubenswrapper[4815]: W1205 09:21:59.162286 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podde650b0f_13f4_4621_9026_d812aec9a752.slice/crio-6d7968c6241f6276b58c522e178f0f9980e3b289a23af9286ae3edbaa35a3763 WatchSource:0}: Error finding container 6d7968c6241f6276b58c522e178f0f9980e3b289a23af9286ae3edbaa35a3763: Status 404 returned error can't find the container with id 6d7968c6241f6276b58c522e178f0f9980e3b289a23af9286ae3edbaa35a3763 Dec 05 09:21:59 crc kubenswrapper[4815]: I1205 09:21:59.468868 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j7bd" event={"ID":"de650b0f-13f4-4621-9026-d812aec9a752","Type":"ContainerStarted","Data":"6d7968c6241f6276b58c522e178f0f9980e3b289a23af9286ae3edbaa35a3763"} Dec 05 09:22:00 crc kubenswrapper[4815]: I1205 09:22:00.475937 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j7bd" event={"ID":"de650b0f-13f4-4621-9026-d812aec9a752","Type":"ContainerStarted","Data":"ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba"} Dec 05 09:22:01 crc kubenswrapper[4815]: I1205 09:22:01.482612 4815 generic.go:334] "Generic (PLEG): container finished" podID="de650b0f-13f4-4621-9026-d812aec9a752" containerID="ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba" exitCode=0 Dec 05 09:22:01 crc kubenswrapper[4815]: I1205 09:22:01.482768 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j7bd" event={"ID":"de650b0f-13f4-4621-9026-d812aec9a752","Type":"ContainerDied","Data":"ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba"} Dec 05 09:22:02 crc kubenswrapper[4815]: I1205 09:22:02.262055 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-69858568cc-scmsp" Dec 05 09:22:03 crc kubenswrapper[4815]: I1205 09:22:03.494517 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j7bd" event={"ID":"de650b0f-13f4-4621-9026-d812aec9a752","Type":"ContainerStarted","Data":"d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802"} Dec 05 09:22:05 crc kubenswrapper[4815]: I1205 09:22:05.510741 4815 generic.go:334] "Generic (PLEG): container finished" podID="de650b0f-13f4-4621-9026-d812aec9a752" containerID="d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802" exitCode=0 Dec 05 09:22:05 crc kubenswrapper[4815]: I1205 09:22:05.510833 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j7bd" event={"ID":"de650b0f-13f4-4621-9026-d812aec9a752","Type":"ContainerDied","Data":"d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802"} Dec 05 09:22:06 crc kubenswrapper[4815]: I1205 09:22:06.521943 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j7bd" event={"ID":"de650b0f-13f4-4621-9026-d812aec9a752","Type":"ContainerStarted","Data":"afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90"} Dec 05 09:22:06 crc kubenswrapper[4815]: I1205 09:22:06.556608 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4j7bd" podStartSLOduration=4.137669715 podStartE2EDuration="8.556565949s" podCreationTimestamp="2025-12-05 09:21:58 +0000 UTC" firstStartedPulling="2025-12-05 09:22:01.484621295 +0000 UTC m=+920.363228132" lastFinishedPulling="2025-12-05 09:22:05.903517529 +0000 UTC m=+924.782124366" observedRunningTime="2025-12-05 09:22:06.552351334 +0000 UTC m=+925.430958171" watchObservedRunningTime="2025-12-05 09:22:06.556565949 +0000 UTC m=+925.435172796" Dec 05 09:22:08 crc kubenswrapper[4815]: I1205 09:22:08.781452 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:22:08 crc kubenswrapper[4815]: I1205 09:22:08.781844 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:22:09 crc kubenswrapper[4815]: I1205 09:22:09.073981 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:22:18 crc kubenswrapper[4815]: I1205 09:22:18.819225 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:22:18 crc kubenswrapper[4815]: I1205 09:22:18.860700 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4j7bd"] Dec 05 09:22:19 crc kubenswrapper[4815]: I1205 09:22:19.658214 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4j7bd" podUID="de650b0f-13f4-4621-9026-d812aec9a752" containerName="registry-server" containerID="cri-o://afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90" gracePeriod=2 Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.192352 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.192750 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.504774 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.665879 4815 generic.go:334] "Generic (PLEG): container finished" podID="de650b0f-13f4-4621-9026-d812aec9a752" containerID="afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90" exitCode=0 Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.665929 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j7bd" event={"ID":"de650b0f-13f4-4621-9026-d812aec9a752","Type":"ContainerDied","Data":"afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90"} Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.666020 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j7bd" event={"ID":"de650b0f-13f4-4621-9026-d812aec9a752","Type":"ContainerDied","Data":"6d7968c6241f6276b58c522e178f0f9980e3b289a23af9286ae3edbaa35a3763"} Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.666045 4815 scope.go:117] "RemoveContainer" containerID="afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.666202 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4j7bd" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.689644 4815 scope.go:117] "RemoveContainer" containerID="d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.700047 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-utilities\") pod \"de650b0f-13f4-4621-9026-d812aec9a752\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.700192 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbs5f\" (UniqueName: \"kubernetes.io/projected/de650b0f-13f4-4621-9026-d812aec9a752-kube-api-access-gbs5f\") pod \"de650b0f-13f4-4621-9026-d812aec9a752\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.700223 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-catalog-content\") pod \"de650b0f-13f4-4621-9026-d812aec9a752\" (UID: \"de650b0f-13f4-4621-9026-d812aec9a752\") " Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.701207 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-utilities" (OuterVolumeSpecName: "utilities") pod "de650b0f-13f4-4621-9026-d812aec9a752" (UID: "de650b0f-13f4-4621-9026-d812aec9a752"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.706278 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de650b0f-13f4-4621-9026-d812aec9a752-kube-api-access-gbs5f" (OuterVolumeSpecName: "kube-api-access-gbs5f") pod "de650b0f-13f4-4621-9026-d812aec9a752" (UID: "de650b0f-13f4-4621-9026-d812aec9a752"). InnerVolumeSpecName "kube-api-access-gbs5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.726452 4815 scope.go:117] "RemoveContainer" containerID="ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.749011 4815 scope.go:117] "RemoveContainer" containerID="afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90" Dec 05 09:22:20 crc kubenswrapper[4815]: E1205 09:22:20.753119 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90\": container with ID starting with afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90 not found: ID does not exist" containerID="afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.753162 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90"} err="failed to get container status \"afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90\": rpc error: code = NotFound desc = could not find container \"afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90\": container with ID starting with afe8b541f7711387db4977f00d87e99620b527c82458cd4f7f0bd598423d7c90 not found: ID does not exist" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.753191 4815 scope.go:117] "RemoveContainer" containerID="d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802" Dec 05 09:22:20 crc kubenswrapper[4815]: E1205 09:22:20.753562 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802\": container with ID starting with d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802 not found: ID does not exist" containerID="d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.753597 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802"} err="failed to get container status \"d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802\": rpc error: code = NotFound desc = could not find container \"d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802\": container with ID starting with d74f6da757fd8bd3909d81f851fa0d755f13ea1b5b6cf5f8adfc8aa74e56f802 not found: ID does not exist" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.753617 4815 scope.go:117] "RemoveContainer" containerID="ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba" Dec 05 09:22:20 crc kubenswrapper[4815]: E1205 09:22:20.753844 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba\": container with ID starting with ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba not found: ID does not exist" containerID="ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.753880 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba"} err="failed to get container status \"ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba\": rpc error: code = NotFound desc = could not find container \"ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba\": container with ID starting with ed2b639c458338d100150eafd3d8ff7007b842eadb29358ec8414de0b6089dba not found: ID does not exist" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.765927 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "de650b0f-13f4-4621-9026-d812aec9a752" (UID: "de650b0f-13f4-4621-9026-d812aec9a752"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.801694 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbs5f\" (UniqueName: \"kubernetes.io/projected/de650b0f-13f4-4621-9026-d812aec9a752-kube-api-access-gbs5f\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.801738 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.801752 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/de650b0f-13f4-4621-9026-d812aec9a752-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.993210 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4j7bd"] Dec 05 09:22:20 crc kubenswrapper[4815]: I1205 09:22:20.999171 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4j7bd"] Dec 05 09:22:21 crc kubenswrapper[4815]: I1205 09:22:21.426639 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de650b0f-13f4-4621-9026-d812aec9a752" path="/var/lib/kubelet/pods/de650b0f-13f4-4621-9026-d812aec9a752/volumes" Dec 05 09:22:21 crc kubenswrapper[4815]: I1205 09:22:21.926196 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-7f5859557b-zxp74" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.886875 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn"] Dec 05 09:22:22 crc kubenswrapper[4815]: E1205 09:22:22.887757 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de650b0f-13f4-4621-9026-d812aec9a752" containerName="extract-content" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.887881 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="de650b0f-13f4-4621-9026-d812aec9a752" containerName="extract-content" Dec 05 09:22:22 crc kubenswrapper[4815]: E1205 09:22:22.887959 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de650b0f-13f4-4621-9026-d812aec9a752" containerName="extract-utilities" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.888027 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="de650b0f-13f4-4621-9026-d812aec9a752" containerName="extract-utilities" Dec 05 09:22:22 crc kubenswrapper[4815]: E1205 09:22:22.888104 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de650b0f-13f4-4621-9026-d812aec9a752" containerName="registry-server" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.888172 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="de650b0f-13f4-4621-9026-d812aec9a752" containerName="registry-server" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.888353 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="de650b0f-13f4-4621-9026-d812aec9a752" containerName="registry-server" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.888941 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.891994 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-hjz5r" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.892258 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.896575 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-jbdfv"] Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.899573 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.901792 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.903581 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 05 09:22:22 crc kubenswrapper[4815]: I1205 09:22:22.943820 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn"] Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.032463 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-frr-conf\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.032523 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-frr-startup\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.032548 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66sx9\" (UniqueName: \"kubernetes.io/projected/46274b62-2000-4c62-8688-af6b5fcb87dc-kube-api-access-66sx9\") pod \"frr-k8s-webhook-server-7fcb986d4-b6cnn\" (UID: \"46274b62-2000-4c62-8688-af6b5fcb87dc\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.032573 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46274b62-2000-4c62-8688-af6b5fcb87dc-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-b6cnn\" (UID: \"46274b62-2000-4c62-8688-af6b5fcb87dc\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.032592 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6hhp\" (UniqueName: \"kubernetes.io/projected/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-kube-api-access-k6hhp\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.032618 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-metrics-certs\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.032647 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-reloader\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.032668 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-metrics\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.032685 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-frr-sockets\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.057451 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-xszxj"] Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.058609 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.062923 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.063217 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.063397 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.068566 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-ltmrh" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.092771 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-j8cbz"] Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.093697 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.096084 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.136128 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-j8cbz"] Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.136163 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-reloader\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.136213 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-metrics\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.136251 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-frr-sockets\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.136291 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-frr-conf\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.136313 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-frr-startup\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.136370 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66sx9\" (UniqueName: \"kubernetes.io/projected/46274b62-2000-4c62-8688-af6b5fcb87dc-kube-api-access-66sx9\") pod \"frr-k8s-webhook-server-7fcb986d4-b6cnn\" (UID: \"46274b62-2000-4c62-8688-af6b5fcb87dc\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.136422 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46274b62-2000-4c62-8688-af6b5fcb87dc-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-b6cnn\" (UID: \"46274b62-2000-4c62-8688-af6b5fcb87dc\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.136446 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6hhp\" (UniqueName: \"kubernetes.io/projected/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-kube-api-access-k6hhp\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.136506 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-metrics-certs\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: E1205 09:22:23.136657 4815 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 05 09:22:23 crc kubenswrapper[4815]: E1205 09:22:23.137055 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-metrics-certs podName:563ef3a0-d97f-4b5b-afc4-378bdbbee19b nodeName:}" failed. No retries permitted until 2025-12-05 09:22:23.637020441 +0000 UTC m=+942.515627278 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-metrics-certs") pod "frr-k8s-jbdfv" (UID: "563ef3a0-d97f-4b5b-afc4-378bdbbee19b") : secret "frr-k8s-certs-secret" not found Dec 05 09:22:23 crc kubenswrapper[4815]: E1205 09:22:23.137617 4815 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 05 09:22:23 crc kubenswrapper[4815]: E1205 09:22:23.137695 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/46274b62-2000-4c62-8688-af6b5fcb87dc-cert podName:46274b62-2000-4c62-8688-af6b5fcb87dc nodeName:}" failed. No retries permitted until 2025-12-05 09:22:23.637666968 +0000 UTC m=+942.516273895 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/46274b62-2000-4c62-8688-af6b5fcb87dc-cert") pod "frr-k8s-webhook-server-7fcb986d4-b6cnn" (UID: "46274b62-2000-4c62-8688-af6b5fcb87dc") : secret "frr-k8s-webhook-server-cert" not found Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.137697 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-frr-startup\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.150896 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-frr-conf\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.151482 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-frr-sockets\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.151777 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-metrics\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.152185 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-reloader\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.191388 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6hhp\" (UniqueName: \"kubernetes.io/projected/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-kube-api-access-k6hhp\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.201257 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66sx9\" (UniqueName: \"kubernetes.io/projected/46274b62-2000-4c62-8688-af6b5fcb87dc-kube-api-access-66sx9\") pod \"frr-k8s-webhook-server-7fcb986d4-b6cnn\" (UID: \"46274b62-2000-4c62-8688-af6b5fcb87dc\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.243296 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-metrics-certs\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.243344 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a72253a0-2394-424d-8689-1c989ce6b2ed-metallb-excludel2\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.243394 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55c34ec0-bd62-4eb5-a838-48e0622ebea9-cert\") pod \"controller-f8648f98b-j8cbz\" (UID: \"55c34ec0-bd62-4eb5-a838-48e0622ebea9\") " pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.243411 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-memberlist\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.243429 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfbkx\" (UniqueName: \"kubernetes.io/projected/a72253a0-2394-424d-8689-1c989ce6b2ed-kube-api-access-pfbkx\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.243457 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jtgm\" (UniqueName: \"kubernetes.io/projected/55c34ec0-bd62-4eb5-a838-48e0622ebea9-kube-api-access-6jtgm\") pod \"controller-f8648f98b-j8cbz\" (UID: \"55c34ec0-bd62-4eb5-a838-48e0622ebea9\") " pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.243500 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55c34ec0-bd62-4eb5-a838-48e0622ebea9-metrics-certs\") pod \"controller-f8648f98b-j8cbz\" (UID: \"55c34ec0-bd62-4eb5-a838-48e0622ebea9\") " pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.344340 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55c34ec0-bd62-4eb5-a838-48e0622ebea9-cert\") pod \"controller-f8648f98b-j8cbz\" (UID: \"55c34ec0-bd62-4eb5-a838-48e0622ebea9\") " pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.344388 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-memberlist\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.344415 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfbkx\" (UniqueName: \"kubernetes.io/projected/a72253a0-2394-424d-8689-1c989ce6b2ed-kube-api-access-pfbkx\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.344449 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jtgm\" (UniqueName: \"kubernetes.io/projected/55c34ec0-bd62-4eb5-a838-48e0622ebea9-kube-api-access-6jtgm\") pod \"controller-f8648f98b-j8cbz\" (UID: \"55c34ec0-bd62-4eb5-a838-48e0622ebea9\") " pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.344506 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55c34ec0-bd62-4eb5-a838-48e0622ebea9-metrics-certs\") pod \"controller-f8648f98b-j8cbz\" (UID: \"55c34ec0-bd62-4eb5-a838-48e0622ebea9\") " pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.344548 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-metrics-certs\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.344565 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a72253a0-2394-424d-8689-1c989ce6b2ed-metallb-excludel2\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: E1205 09:22:23.345079 4815 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 05 09:22:23 crc kubenswrapper[4815]: E1205 09:22:23.345195 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-memberlist podName:a72253a0-2394-424d-8689-1c989ce6b2ed nodeName:}" failed. No retries permitted until 2025-12-05 09:22:23.84517911 +0000 UTC m=+942.723785947 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-memberlist") pod "speaker-xszxj" (UID: "a72253a0-2394-424d-8689-1c989ce6b2ed") : secret "metallb-memberlist" not found Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.345220 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/a72253a0-2394-424d-8689-1c989ce6b2ed-metallb-excludel2\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: E1205 09:22:23.345143 4815 secret.go:188] Couldn't get secret metallb-system/controller-certs-secret: secret "controller-certs-secret" not found Dec 05 09:22:23 crc kubenswrapper[4815]: E1205 09:22:23.345421 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55c34ec0-bd62-4eb5-a838-48e0622ebea9-metrics-certs podName:55c34ec0-bd62-4eb5-a838-48e0622ebea9 nodeName:}" failed. No retries permitted until 2025-12-05 09:22:23.845408726 +0000 UTC m=+942.724015563 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/55c34ec0-bd62-4eb5-a838-48e0622ebea9-metrics-certs") pod "controller-f8648f98b-j8cbz" (UID: "55c34ec0-bd62-4eb5-a838-48e0622ebea9") : secret "controller-certs-secret" not found Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.349005 4815 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.349142 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-metrics-certs\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.358326 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55c34ec0-bd62-4eb5-a838-48e0622ebea9-cert\") pod \"controller-f8648f98b-j8cbz\" (UID: \"55c34ec0-bd62-4eb5-a838-48e0622ebea9\") " pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.376575 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jtgm\" (UniqueName: \"kubernetes.io/projected/55c34ec0-bd62-4eb5-a838-48e0622ebea9-kube-api-access-6jtgm\") pod \"controller-f8648f98b-j8cbz\" (UID: \"55c34ec0-bd62-4eb5-a838-48e0622ebea9\") " pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.379939 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfbkx\" (UniqueName: \"kubernetes.io/projected/a72253a0-2394-424d-8689-1c989ce6b2ed-kube-api-access-pfbkx\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.647421 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-metrics-certs\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.647582 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46274b62-2000-4c62-8688-af6b5fcb87dc-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-b6cnn\" (UID: \"46274b62-2000-4c62-8688-af6b5fcb87dc\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.654036 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/563ef3a0-d97f-4b5b-afc4-378bdbbee19b-metrics-certs\") pod \"frr-k8s-jbdfv\" (UID: \"563ef3a0-d97f-4b5b-afc4-378bdbbee19b\") " pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.654196 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/46274b62-2000-4c62-8688-af6b5fcb87dc-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-b6cnn\" (UID: \"46274b62-2000-4c62-8688-af6b5fcb87dc\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.811848 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.827664 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.849649 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-memberlist\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.849738 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55c34ec0-bd62-4eb5-a838-48e0622ebea9-metrics-certs\") pod \"controller-f8648f98b-j8cbz\" (UID: \"55c34ec0-bd62-4eb5-a838-48e0622ebea9\") " pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:23 crc kubenswrapper[4815]: E1205 09:22:23.849920 4815 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 05 09:22:23 crc kubenswrapper[4815]: E1205 09:22:23.850009 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-memberlist podName:a72253a0-2394-424d-8689-1c989ce6b2ed nodeName:}" failed. No retries permitted until 2025-12-05 09:22:24.849989292 +0000 UTC m=+943.728596159 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-memberlist") pod "speaker-xszxj" (UID: "a72253a0-2394-424d-8689-1c989ce6b2ed") : secret "metallb-memberlist" not found Dec 05 09:22:23 crc kubenswrapper[4815]: I1205 09:22:23.855648 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/55c34ec0-bd62-4eb5-a838-48e0622ebea9-metrics-certs\") pod \"controller-f8648f98b-j8cbz\" (UID: \"55c34ec0-bd62-4eb5-a838-48e0622ebea9\") " pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.005158 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.254228 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn"] Dec 05 09:22:24 crc kubenswrapper[4815]: W1205 09:22:24.259535 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46274b62_2000_4c62_8688_af6b5fcb87dc.slice/crio-55a03316bb05f018c252e5086ba6e221c797af8d7fc64691bf9cd6d12e00aef8 WatchSource:0}: Error finding container 55a03316bb05f018c252e5086ba6e221c797af8d7fc64691bf9cd6d12e00aef8: Status 404 returned error can't find the container with id 55a03316bb05f018c252e5086ba6e221c797af8d7fc64691bf9cd6d12e00aef8 Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.288351 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-j8cbz"] Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.394570 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gkvsp"] Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.397220 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.404217 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkvsp"] Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.564419 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dhq6\" (UniqueName: \"kubernetes.io/projected/24cd25c3-6c61-4d4e-8af5-26fefc5405da-kube-api-access-6dhq6\") pod \"redhat-marketplace-gkvsp\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.564850 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-utilities\") pod \"redhat-marketplace-gkvsp\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.564881 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-catalog-content\") pod \"redhat-marketplace-gkvsp\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.666235 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dhq6\" (UniqueName: \"kubernetes.io/projected/24cd25c3-6c61-4d4e-8af5-26fefc5405da-kube-api-access-6dhq6\") pod \"redhat-marketplace-gkvsp\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.666283 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-utilities\") pod \"redhat-marketplace-gkvsp\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.666321 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-catalog-content\") pod \"redhat-marketplace-gkvsp\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.666845 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-catalog-content\") pod \"redhat-marketplace-gkvsp\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.667148 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-utilities\") pod \"redhat-marketplace-gkvsp\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.687927 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dhq6\" (UniqueName: \"kubernetes.io/projected/24cd25c3-6c61-4d4e-8af5-26fefc5405da-kube-api-access-6dhq6\") pod \"redhat-marketplace-gkvsp\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.700192 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-j8cbz" event={"ID":"55c34ec0-bd62-4eb5-a838-48e0622ebea9","Type":"ContainerStarted","Data":"de114ca04f0a5383522232cad790e380b16a7429aa57c1a12d737b4c66c85520"} Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.700236 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-j8cbz" event={"ID":"55c34ec0-bd62-4eb5-a838-48e0622ebea9","Type":"ContainerStarted","Data":"b0a318fd91840af84e730c6243c34e9ffd25e2740ee5426374f20edf16e905f1"} Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.702591 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbdfv" event={"ID":"563ef3a0-d97f-4b5b-afc4-378bdbbee19b","Type":"ContainerStarted","Data":"448002ac05a8ff66051638316155994aca74e04a9ee28bf04bbbce741de73264"} Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.708595 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" event={"ID":"46274b62-2000-4c62-8688-af6b5fcb87dc","Type":"ContainerStarted","Data":"55a03316bb05f018c252e5086ba6e221c797af8d7fc64691bf9cd6d12e00aef8"} Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.739617 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.870295 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-memberlist\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:24 crc kubenswrapper[4815]: I1205 09:22:24.876168 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/a72253a0-2394-424d-8689-1c989ce6b2ed-memberlist\") pod \"speaker-xszxj\" (UID: \"a72253a0-2394-424d-8689-1c989ce6b2ed\") " pod="metallb-system/speaker-xszxj" Dec 05 09:22:25 crc kubenswrapper[4815]: I1205 09:22:25.050286 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkvsp"] Dec 05 09:22:25 crc kubenswrapper[4815]: W1205 09:22:25.058552 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24cd25c3_6c61_4d4e_8af5_26fefc5405da.slice/crio-e96e41bc7eb6e8030e2562bd929e0676848ae5c47d9e0accfc2e559217107ca9 WatchSource:0}: Error finding container e96e41bc7eb6e8030e2562bd929e0676848ae5c47d9e0accfc2e559217107ca9: Status 404 returned error can't find the container with id e96e41bc7eb6e8030e2562bd929e0676848ae5c47d9e0accfc2e559217107ca9 Dec 05 09:22:25 crc kubenswrapper[4815]: I1205 09:22:25.172578 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-xszxj" Dec 05 09:22:25 crc kubenswrapper[4815]: W1205 09:22:25.194053 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda72253a0_2394_424d_8689_1c989ce6b2ed.slice/crio-b1a488782cf86c9cfe504faccce2bd3f74a3bfcf3f03f5331725f2b9402eeeb2 WatchSource:0}: Error finding container b1a488782cf86c9cfe504faccce2bd3f74a3bfcf3f03f5331725f2b9402eeeb2: Status 404 returned error can't find the container with id b1a488782cf86c9cfe504faccce2bd3f74a3bfcf3f03f5331725f2b9402eeeb2 Dec 05 09:22:25 crc kubenswrapper[4815]: I1205 09:22:25.716539 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkvsp" event={"ID":"24cd25c3-6c61-4d4e-8af5-26fefc5405da","Type":"ContainerStarted","Data":"cab22efa889cc1e126df723ed9efdf1259d61e9fcb06f1d71fd9b9296a6cf87f"} Dec 05 09:22:25 crc kubenswrapper[4815]: I1205 09:22:25.716768 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkvsp" event={"ID":"24cd25c3-6c61-4d4e-8af5-26fefc5405da","Type":"ContainerStarted","Data":"e96e41bc7eb6e8030e2562bd929e0676848ae5c47d9e0accfc2e559217107ca9"} Dec 05 09:22:25 crc kubenswrapper[4815]: I1205 09:22:25.718163 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-j8cbz" event={"ID":"55c34ec0-bd62-4eb5-a838-48e0622ebea9","Type":"ContainerStarted","Data":"4401e75a020e46b351479e922d9274a25f7a1f92edf06e00ac1bbf1e57ddaa96"} Dec 05 09:22:25 crc kubenswrapper[4815]: I1205 09:22:25.718284 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:25 crc kubenswrapper[4815]: I1205 09:22:25.719481 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xszxj" event={"ID":"a72253a0-2394-424d-8689-1c989ce6b2ed","Type":"ContainerStarted","Data":"b1a488782cf86c9cfe504faccce2bd3f74a3bfcf3f03f5331725f2b9402eeeb2"} Dec 05 09:22:25 crc kubenswrapper[4815]: I1205 09:22:25.742811 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-j8cbz" podStartSLOduration=2.742792814 podStartE2EDuration="2.742792814s" podCreationTimestamp="2025-12-05 09:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:22:25.741898279 +0000 UTC m=+944.620505116" watchObservedRunningTime="2025-12-05 09:22:25.742792814 +0000 UTC m=+944.621399651" Dec 05 09:22:26 crc kubenswrapper[4815]: I1205 09:22:26.787561 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xszxj" event={"ID":"a72253a0-2394-424d-8689-1c989ce6b2ed","Type":"ContainerStarted","Data":"227b0b4826ea86e68a365e892b4b1926e1c552de12bdbc9d1239364e1089d041"} Dec 05 09:22:26 crc kubenswrapper[4815]: I1205 09:22:26.788676 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-xszxj" event={"ID":"a72253a0-2394-424d-8689-1c989ce6b2ed","Type":"ContainerStarted","Data":"fb320e0f7adeb06124298601333036c3c7ed1b505791f74e4808f86104728013"} Dec 05 09:22:26 crc kubenswrapper[4815]: I1205 09:22:26.794056 4815 generic.go:334] "Generic (PLEG): container finished" podID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" containerID="cab22efa889cc1e126df723ed9efdf1259d61e9fcb06f1d71fd9b9296a6cf87f" exitCode=0 Dec 05 09:22:26 crc kubenswrapper[4815]: I1205 09:22:26.794870 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkvsp" event={"ID":"24cd25c3-6c61-4d4e-8af5-26fefc5405da","Type":"ContainerDied","Data":"cab22efa889cc1e126df723ed9efdf1259d61e9fcb06f1d71fd9b9296a6cf87f"} Dec 05 09:22:27 crc kubenswrapper[4815]: I1205 09:22:27.803073 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-xszxj" Dec 05 09:22:27 crc kubenswrapper[4815]: I1205 09:22:27.826537 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-xszxj" podStartSLOduration=4.826516204 podStartE2EDuration="4.826516204s" podCreationTimestamp="2025-12-05 09:22:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:22:27.823037289 +0000 UTC m=+946.701644136" watchObservedRunningTime="2025-12-05 09:22:27.826516204 +0000 UTC m=+946.705123051" Dec 05 09:22:28 crc kubenswrapper[4815]: I1205 09:22:28.899775 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hx6gq"] Dec 05 09:22:28 crc kubenswrapper[4815]: I1205 09:22:28.976902 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkvsp" event={"ID":"24cd25c3-6c61-4d4e-8af5-26fefc5405da","Type":"ContainerStarted","Data":"5c2ac6e796278a0dce69a5fccbeaaff2b0b9be9368c554fb53f8e98600654ba4"} Dec 05 09:22:28 crc kubenswrapper[4815]: I1205 09:22:28.977192 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hx6gq"] Dec 05 09:22:28 crc kubenswrapper[4815]: I1205 09:22:28.977330 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.106524 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-utilities\") pod \"community-operators-hx6gq\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.106560 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-catalog-content\") pod \"community-operators-hx6gq\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.106619 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8jrq\" (UniqueName: \"kubernetes.io/projected/35a4b760-e48f-428b-8ec9-8779a968af92-kube-api-access-d8jrq\") pod \"community-operators-hx6gq\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.207312 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-utilities\") pod \"community-operators-hx6gq\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.207367 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-catalog-content\") pod \"community-operators-hx6gq\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.207409 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8jrq\" (UniqueName: \"kubernetes.io/projected/35a4b760-e48f-428b-8ec9-8779a968af92-kube-api-access-d8jrq\") pod \"community-operators-hx6gq\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.207830 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-utilities\") pod \"community-operators-hx6gq\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.208091 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-catalog-content\") pod \"community-operators-hx6gq\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.231411 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8jrq\" (UniqueName: \"kubernetes.io/projected/35a4b760-e48f-428b-8ec9-8779a968af92-kube-api-access-d8jrq\") pod \"community-operators-hx6gq\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.336125 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.968947 4815 generic.go:334] "Generic (PLEG): container finished" podID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" containerID="5c2ac6e796278a0dce69a5fccbeaaff2b0b9be9368c554fb53f8e98600654ba4" exitCode=0 Dec 05 09:22:29 crc kubenswrapper[4815]: I1205 09:22:29.968995 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkvsp" event={"ID":"24cd25c3-6c61-4d4e-8af5-26fefc5405da","Type":"ContainerDied","Data":"5c2ac6e796278a0dce69a5fccbeaaff2b0b9be9368c554fb53f8e98600654ba4"} Dec 05 09:22:30 crc kubenswrapper[4815]: I1205 09:22:30.150472 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hx6gq"] Dec 05 09:22:30 crc kubenswrapper[4815]: W1205 09:22:30.178794 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35a4b760_e48f_428b_8ec9_8779a968af92.slice/crio-7a97af43e3f0d05fe20670a742a9de1217241423978b849048a58d9ce4e575d5 WatchSource:0}: Error finding container 7a97af43e3f0d05fe20670a742a9de1217241423978b849048a58d9ce4e575d5: Status 404 returned error can't find the container with id 7a97af43e3f0d05fe20670a742a9de1217241423978b849048a58d9ce4e575d5 Dec 05 09:22:30 crc kubenswrapper[4815]: I1205 09:22:30.976574 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx6gq" event={"ID":"35a4b760-e48f-428b-8ec9-8779a968af92","Type":"ContainerStarted","Data":"7a97af43e3f0d05fe20670a742a9de1217241423978b849048a58d9ce4e575d5"} Dec 05 09:22:34 crc kubenswrapper[4815]: I1205 09:22:34.010442 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-j8cbz" Dec 05 09:22:35 crc kubenswrapper[4815]: I1205 09:22:35.178900 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-xszxj" Dec 05 09:22:37 crc kubenswrapper[4815]: I1205 09:22:37.065383 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx6gq" event={"ID":"35a4b760-e48f-428b-8ec9-8779a968af92","Type":"ContainerStarted","Data":"18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f"} Dec 05 09:22:38 crc kubenswrapper[4815]: I1205 09:22:38.073942 4815 generic.go:334] "Generic (PLEG): container finished" podID="35a4b760-e48f-428b-8ec9-8779a968af92" containerID="18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f" exitCode=0 Dec 05 09:22:38 crc kubenswrapper[4815]: I1205 09:22:38.073995 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx6gq" event={"ID":"35a4b760-e48f-428b-8ec9-8779a968af92","Type":"ContainerDied","Data":"18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f"} Dec 05 09:22:38 crc kubenswrapper[4815]: I1205 09:22:38.841009 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xttqw"] Dec 05 09:22:38 crc kubenswrapper[4815]: I1205 09:22:38.842271 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xttqw" Dec 05 09:22:38 crc kubenswrapper[4815]: I1205 09:22:38.845657 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-2pscm" Dec 05 09:22:38 crc kubenswrapper[4815]: I1205 09:22:38.845780 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 05 09:22:38 crc kubenswrapper[4815]: I1205 09:22:38.847480 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 05 09:22:38 crc kubenswrapper[4815]: I1205 09:22:38.862328 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xttqw"] Dec 05 09:22:38 crc kubenswrapper[4815]: I1205 09:22:38.903750 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b9lx6\" (UniqueName: \"kubernetes.io/projected/6943ef97-36ce-4f74-b31b-4ba90192d3ac-kube-api-access-b9lx6\") pod \"openstack-operator-index-xttqw\" (UID: \"6943ef97-36ce-4f74-b31b-4ba90192d3ac\") " pod="openstack-operators/openstack-operator-index-xttqw" Dec 05 09:22:39 crc kubenswrapper[4815]: I1205 09:22:39.005235 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b9lx6\" (UniqueName: \"kubernetes.io/projected/6943ef97-36ce-4f74-b31b-4ba90192d3ac-kube-api-access-b9lx6\") pod \"openstack-operator-index-xttqw\" (UID: \"6943ef97-36ce-4f74-b31b-4ba90192d3ac\") " pod="openstack-operators/openstack-operator-index-xttqw" Dec 05 09:22:39 crc kubenswrapper[4815]: I1205 09:22:39.024050 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b9lx6\" (UniqueName: \"kubernetes.io/projected/6943ef97-36ce-4f74-b31b-4ba90192d3ac-kube-api-access-b9lx6\") pod \"openstack-operator-index-xttqw\" (UID: \"6943ef97-36ce-4f74-b31b-4ba90192d3ac\") " pod="openstack-operators/openstack-operator-index-xttqw" Dec 05 09:22:39 crc kubenswrapper[4815]: I1205 09:22:39.163562 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xttqw" Dec 05 09:22:44 crc kubenswrapper[4815]: E1205 09:22:44.045535 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a" Dec 05 09:22:44 crc kubenswrapper[4815]: E1205 09:22:44.046308 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:frr-k8s-webhook-server,Image:registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a,Command:[/frr-k8s],Args:[--log-level=debug --webhook-mode=onlywebhook --disable-cert-rotation=true --namespace=$(NAMESPACE) --metrics-bind-address=:7572],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:monitoring,HostPort:0,ContainerPort:7572,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-66sx9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/metrics,Port:{1 0 monitoring},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/metrics,Port:{1 0 monitoring},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000700000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod frr-k8s-webhook-server-7fcb986d4-b6cnn_metallb-system(46274b62-2000-4c62-8688-af6b5fcb87dc): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:22:44 crc kubenswrapper[4815]: E1205 09:22:44.047511 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"frr-k8s-webhook-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" podUID="46274b62-2000-4c62-8688-af6b5fcb87dc" Dec 05 09:22:44 crc kubenswrapper[4815]: E1205 09:22:44.272745 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"frr-k8s-webhook-server\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a\\\"\"" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" podUID="46274b62-2000-4c62-8688-af6b5fcb87dc" Dec 05 09:22:44 crc kubenswrapper[4815]: E1205 09:22:44.316717 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a" Dec 05 09:22:44 crc kubenswrapper[4815]: E1205 09:22:44.316907 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:cp-frr-files,Image:registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a,Command:[/bin/sh -c cp -rLf /tmp/frr/* /etc/frr/],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:frr-startup,ReadOnly:false,MountPath:/tmp/frr,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:frr-conf,ReadOnly:false,MountPath:/etc/frr,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k6hhp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*100,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*101,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod frr-k8s-jbdfv_metallb-system(563ef3a0-d97f-4b5b-afc4-378bdbbee19b): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:22:44 crc kubenswrapper[4815]: E1205 09:22:44.318731 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cp-frr-files\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="metallb-system/frr-k8s-jbdfv" podUID="563ef3a0-d97f-4b5b-afc4-378bdbbee19b" Dec 05 09:22:44 crc kubenswrapper[4815]: I1205 09:22:44.726777 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xttqw"] Dec 05 09:22:44 crc kubenswrapper[4815]: W1205 09:22:44.737384 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6943ef97_36ce_4f74_b31b_4ba90192d3ac.slice/crio-d3b984be5440a33e03fce83efc33c9441f0ddd6a20d71eef56dd8768a1958f40 WatchSource:0}: Error finding container d3b984be5440a33e03fce83efc33c9441f0ddd6a20d71eef56dd8768a1958f40: Status 404 returned error can't find the container with id d3b984be5440a33e03fce83efc33c9441f0ddd6a20d71eef56dd8768a1958f40 Dec 05 09:22:45 crc kubenswrapper[4815]: I1205 09:22:45.202827 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xttqw" event={"ID":"6943ef97-36ce-4f74-b31b-4ba90192d3ac","Type":"ContainerStarted","Data":"d3b984be5440a33e03fce83efc33c9441f0ddd6a20d71eef56dd8768a1958f40"} Dec 05 09:22:45 crc kubenswrapper[4815]: I1205 09:22:45.205352 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkvsp" event={"ID":"24cd25c3-6c61-4d4e-8af5-26fefc5405da","Type":"ContainerStarted","Data":"82e4090577f3730ae6988204964f52ee04760bb7ba8a02bec2e575c1903f51b7"} Dec 05 09:22:45 crc kubenswrapper[4815]: I1205 09:22:45.224836 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gkvsp" podStartSLOduration=3.7490606890000002 podStartE2EDuration="21.224817787s" podCreationTimestamp="2025-12-05 09:22:24 +0000 UTC" firstStartedPulling="2025-12-05 09:22:26.796549351 +0000 UTC m=+945.675156188" lastFinishedPulling="2025-12-05 09:22:44.272306449 +0000 UTC m=+963.150913286" observedRunningTime="2025-12-05 09:22:45.220387146 +0000 UTC m=+964.098993993" watchObservedRunningTime="2025-12-05 09:22:45.224817787 +0000 UTC m=+964.103424624" Dec 05 09:22:45 crc kubenswrapper[4815]: E1205 09:22:45.312271 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cp-frr-files\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/frr-rhel9@sha256:e5c5e7ca4ed54c9edba5dfa1d504bbe58016c2abdc872ebb8b26a628958e5a2a\\\"\"" pod="metallb-system/frr-k8s-jbdfv" podUID="563ef3a0-d97f-4b5b-afc4-378bdbbee19b" Dec 05 09:22:46 crc kubenswrapper[4815]: I1205 09:22:46.211641 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx6gq" event={"ID":"35a4b760-e48f-428b-8ec9-8779a968af92","Type":"ContainerStarted","Data":"ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd"} Dec 05 09:22:47 crc kubenswrapper[4815]: I1205 09:22:47.219455 4815 generic.go:334] "Generic (PLEG): container finished" podID="35a4b760-e48f-428b-8ec9-8779a968af92" containerID="ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd" exitCode=0 Dec 05 09:22:47 crc kubenswrapper[4815]: I1205 09:22:47.219509 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx6gq" event={"ID":"35a4b760-e48f-428b-8ec9-8779a968af92","Type":"ContainerDied","Data":"ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd"} Dec 05 09:22:49 crc kubenswrapper[4815]: I1205 09:22:49.231483 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xttqw" event={"ID":"6943ef97-36ce-4f74-b31b-4ba90192d3ac","Type":"ContainerStarted","Data":"985912313c35de8706ad820e6acda265c23ccb70590c4c185a5c260e812974fd"} Dec 05 09:22:49 crc kubenswrapper[4815]: I1205 09:22:49.255879 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xttqw" podStartSLOduration=7.752914275 podStartE2EDuration="11.255856723s" podCreationTimestamp="2025-12-05 09:22:38 +0000 UTC" firstStartedPulling="2025-12-05 09:22:44.739140869 +0000 UTC m=+963.617747696" lastFinishedPulling="2025-12-05 09:22:48.242083307 +0000 UTC m=+967.120690144" observedRunningTime="2025-12-05 09:22:49.252799318 +0000 UTC m=+968.131406155" watchObservedRunningTime="2025-12-05 09:22:49.255856723 +0000 UTC m=+968.134463560" Dec 05 09:22:50 crc kubenswrapper[4815]: I1205 09:22:50.192798 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:22:50 crc kubenswrapper[4815]: I1205 09:22:50.192857 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:22:50 crc kubenswrapper[4815]: I1205 09:22:50.240219 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx6gq" event={"ID":"35a4b760-e48f-428b-8ec9-8779a968af92","Type":"ContainerStarted","Data":"7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0"} Dec 05 09:22:50 crc kubenswrapper[4815]: I1205 09:22:50.258088 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hx6gq" podStartSLOduration=17.604518552000002 podStartE2EDuration="22.258067431s" podCreationTimestamp="2025-12-05 09:22:28 +0000 UTC" firstStartedPulling="2025-12-05 09:22:43.867773342 +0000 UTC m=+962.746380179" lastFinishedPulling="2025-12-05 09:22:48.521322221 +0000 UTC m=+967.399929058" observedRunningTime="2025-12-05 09:22:50.255903203 +0000 UTC m=+969.134510040" watchObservedRunningTime="2025-12-05 09:22:50.258067431 +0000 UTC m=+969.136674268" Dec 05 09:22:54 crc kubenswrapper[4815]: I1205 09:22:54.739800 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:54 crc kubenswrapper[4815]: I1205 09:22:54.740102 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:54 crc kubenswrapper[4815]: I1205 09:22:54.782892 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:55 crc kubenswrapper[4815]: I1205 09:22:55.315258 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:57 crc kubenswrapper[4815]: I1205 09:22:57.703767 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkvsp"] Dec 05 09:22:57 crc kubenswrapper[4815]: I1205 09:22:57.703984 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-gkvsp" podUID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" containerName="registry-server" containerID="cri-o://82e4090577f3730ae6988204964f52ee04760bb7ba8a02bec2e575c1903f51b7" gracePeriod=2 Dec 05 09:22:58 crc kubenswrapper[4815]: I1205 09:22:58.290527 4815 generic.go:334] "Generic (PLEG): container finished" podID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" containerID="82e4090577f3730ae6988204964f52ee04760bb7ba8a02bec2e575c1903f51b7" exitCode=0 Dec 05 09:22:58 crc kubenswrapper[4815]: I1205 09:22:58.290818 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkvsp" event={"ID":"24cd25c3-6c61-4d4e-8af5-26fefc5405da","Type":"ContainerDied","Data":"82e4090577f3730ae6988204964f52ee04760bb7ba8a02bec2e575c1903f51b7"} Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.126766 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.164223 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-xttqw" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.164688 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-xttqw" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.204141 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-xttqw" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.248615 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-utilities\") pod \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.248709 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dhq6\" (UniqueName: \"kubernetes.io/projected/24cd25c3-6c61-4d4e-8af5-26fefc5405da-kube-api-access-6dhq6\") pod \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.248788 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-catalog-content\") pod \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\" (UID: \"24cd25c3-6c61-4d4e-8af5-26fefc5405da\") " Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.249718 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-utilities" (OuterVolumeSpecName: "utilities") pod "24cd25c3-6c61-4d4e-8af5-26fefc5405da" (UID: "24cd25c3-6c61-4d4e-8af5-26fefc5405da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.255698 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24cd25c3-6c61-4d4e-8af5-26fefc5405da-kube-api-access-6dhq6" (OuterVolumeSpecName: "kube-api-access-6dhq6") pod "24cd25c3-6c61-4d4e-8af5-26fefc5405da" (UID: "24cd25c3-6c61-4d4e-8af5-26fefc5405da"). InnerVolumeSpecName "kube-api-access-6dhq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.268722 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24cd25c3-6c61-4d4e-8af5-26fefc5405da" (UID: "24cd25c3-6c61-4d4e-8af5-26fefc5405da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.302865 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gkvsp" event={"ID":"24cd25c3-6c61-4d4e-8af5-26fefc5405da","Type":"ContainerDied","Data":"e96e41bc7eb6e8030e2562bd929e0676848ae5c47d9e0accfc2e559217107ca9"} Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.302887 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gkvsp" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.303219 4815 scope.go:117] "RemoveContainer" containerID="82e4090577f3730ae6988204964f52ee04760bb7ba8a02bec2e575c1903f51b7" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.335320 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-xttqw" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.336508 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.336535 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.340034 4815 scope.go:117] "RemoveContainer" containerID="5c2ac6e796278a0dce69a5fccbeaaff2b0b9be9368c554fb53f8e98600654ba4" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.350187 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dhq6\" (UniqueName: \"kubernetes.io/projected/24cd25c3-6c61-4d4e-8af5-26fefc5405da-kube-api-access-6dhq6\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.350219 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.350229 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24cd25c3-6c61-4d4e-8af5-26fefc5405da-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.362306 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkvsp"] Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.388141 4815 scope.go:117] "RemoveContainer" containerID="cab22efa889cc1e126df723ed9efdf1259d61e9fcb06f1d71fd9b9296a6cf87f" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.388521 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-gkvsp"] Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.435112 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" path="/var/lib/kubelet/pods/24cd25c3-6c61-4d4e-8af5-26fefc5405da/volumes" Dec 05 09:22:59 crc kubenswrapper[4815]: I1205 09:22:59.436850 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:23:00 crc kubenswrapper[4815]: I1205 09:23:00.310825 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" event={"ID":"46274b62-2000-4c62-8688-af6b5fcb87dc","Type":"ContainerStarted","Data":"765331c9f9760daa4c192c50db7473b358e10db9019064d861c02e6b0a0d429b"} Dec 05 09:23:00 crc kubenswrapper[4815]: I1205 09:23:00.312554 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" Dec 05 09:23:00 crc kubenswrapper[4815]: I1205 09:23:00.330376 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" podStartSLOduration=3.360069787 podStartE2EDuration="38.33036289s" podCreationTimestamp="2025-12-05 09:22:22 +0000 UTC" firstStartedPulling="2025-12-05 09:22:24.261429956 +0000 UTC m=+943.140036793" lastFinishedPulling="2025-12-05 09:22:59.231723059 +0000 UTC m=+978.110329896" observedRunningTime="2025-12-05 09:23:00.32780859 +0000 UTC m=+979.206415427" watchObservedRunningTime="2025-12-05 09:23:00.33036289 +0000 UTC m=+979.208969727" Dec 05 09:23:00 crc kubenswrapper[4815]: I1205 09:23:00.355466 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:23:01 crc kubenswrapper[4815]: I1205 09:23:01.328381 4815 generic.go:334] "Generic (PLEG): container finished" podID="563ef3a0-d97f-4b5b-afc4-378bdbbee19b" containerID="7a582347148b281d337b89b4455442d126a3c51104578b5a46f9667fee573f00" exitCode=0 Dec 05 09:23:01 crc kubenswrapper[4815]: I1205 09:23:01.328485 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbdfv" event={"ID":"563ef3a0-d97f-4b5b-afc4-378bdbbee19b","Type":"ContainerDied","Data":"7a582347148b281d337b89b4455442d126a3c51104578b5a46f9667fee573f00"} Dec 05 09:23:02 crc kubenswrapper[4815]: I1205 09:23:02.304395 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hx6gq"] Dec 05 09:23:02 crc kubenswrapper[4815]: I1205 09:23:02.338154 4815 generic.go:334] "Generic (PLEG): container finished" podID="563ef3a0-d97f-4b5b-afc4-378bdbbee19b" containerID="b8dc54f4e2c27c0aac0af7dac062b4e2c93932c673baea3d6a19e48b8a1efac4" exitCode=0 Dec 05 09:23:02 crc kubenswrapper[4815]: I1205 09:23:02.338218 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbdfv" event={"ID":"563ef3a0-d97f-4b5b-afc4-378bdbbee19b","Type":"ContainerDied","Data":"b8dc54f4e2c27c0aac0af7dac062b4e2c93932c673baea3d6a19e48b8a1efac4"} Dec 05 09:23:02 crc kubenswrapper[4815]: I1205 09:23:02.339144 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hx6gq" podUID="35a4b760-e48f-428b-8ec9-8779a968af92" containerName="registry-server" containerID="cri-o://7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0" gracePeriod=2 Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.325258 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.356529 4815 generic.go:334] "Generic (PLEG): container finished" podID="563ef3a0-d97f-4b5b-afc4-378bdbbee19b" containerID="6083f5a04c30d97ec1bede321726aa22d1d11e42ce94c6a47675b381022abbf6" exitCode=0 Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.356620 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbdfv" event={"ID":"563ef3a0-d97f-4b5b-afc4-378bdbbee19b","Type":"ContainerDied","Data":"6083f5a04c30d97ec1bede321726aa22d1d11e42ce94c6a47675b381022abbf6"} Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.364957 4815 generic.go:334] "Generic (PLEG): container finished" podID="35a4b760-e48f-428b-8ec9-8779a968af92" containerID="7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0" exitCode=0 Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.364994 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx6gq" event={"ID":"35a4b760-e48f-428b-8ec9-8779a968af92","Type":"ContainerDied","Data":"7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0"} Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.365017 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hx6gq" event={"ID":"35a4b760-e48f-428b-8ec9-8779a968af92","Type":"ContainerDied","Data":"7a97af43e3f0d05fe20670a742a9de1217241423978b849048a58d9ce4e575d5"} Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.365033 4815 scope.go:117] "RemoveContainer" containerID="7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.365118 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hx6gq" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.382685 4815 scope.go:117] "RemoveContainer" containerID="ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.428016 4815 scope.go:117] "RemoveContainer" containerID="18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.471030 4815 scope.go:117] "RemoveContainer" containerID="7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0" Dec 05 09:23:03 crc kubenswrapper[4815]: E1205 09:23:03.472739 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0\": container with ID starting with 7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0 not found: ID does not exist" containerID="7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.472812 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0"} err="failed to get container status \"7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0\": rpc error: code = NotFound desc = could not find container \"7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0\": container with ID starting with 7048f7086a416f76b8a51dd46259a20666f900bc9fbed1bf6a5cdf9d550448c0 not found: ID does not exist" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.472848 4815 scope.go:117] "RemoveContainer" containerID="ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd" Dec 05 09:23:03 crc kubenswrapper[4815]: E1205 09:23:03.473216 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd\": container with ID starting with ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd not found: ID does not exist" containerID="ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.473252 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd"} err="failed to get container status \"ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd\": rpc error: code = NotFound desc = could not find container \"ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd\": container with ID starting with ee62d1f3d0327a392c7e3c7a3441b27e577924916ca60e6116647a8c940aacbd not found: ID does not exist" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.473275 4815 scope.go:117] "RemoveContainer" containerID="18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f" Dec 05 09:23:03 crc kubenswrapper[4815]: E1205 09:23:03.473621 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f\": container with ID starting with 18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f not found: ID does not exist" containerID="18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.473649 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f"} err="failed to get container status \"18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f\": rpc error: code = NotFound desc = could not find container \"18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f\": container with ID starting with 18278107519eca5ac1af3491bf206746d1d41307f627df5b3131711b2da7dc8f not found: ID does not exist" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.506065 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8jrq\" (UniqueName: \"kubernetes.io/projected/35a4b760-e48f-428b-8ec9-8779a968af92-kube-api-access-d8jrq\") pod \"35a4b760-e48f-428b-8ec9-8779a968af92\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.506151 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-catalog-content\") pod \"35a4b760-e48f-428b-8ec9-8779a968af92\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.506190 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-utilities\") pod \"35a4b760-e48f-428b-8ec9-8779a968af92\" (UID: \"35a4b760-e48f-428b-8ec9-8779a968af92\") " Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.508221 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-utilities" (OuterVolumeSpecName: "utilities") pod "35a4b760-e48f-428b-8ec9-8779a968af92" (UID: "35a4b760-e48f-428b-8ec9-8779a968af92"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.512109 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35a4b760-e48f-428b-8ec9-8779a968af92-kube-api-access-d8jrq" (OuterVolumeSpecName: "kube-api-access-d8jrq") pod "35a4b760-e48f-428b-8ec9-8779a968af92" (UID: "35a4b760-e48f-428b-8ec9-8779a968af92"). InnerVolumeSpecName "kube-api-access-d8jrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.556856 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35a4b760-e48f-428b-8ec9-8779a968af92" (UID: "35a4b760-e48f-428b-8ec9-8779a968af92"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.607603 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.608341 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35a4b760-e48f-428b-8ec9-8779a968af92-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.608528 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8jrq\" (UniqueName: \"kubernetes.io/projected/35a4b760-e48f-428b-8ec9-8779a968af92-kube-api-access-d8jrq\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.709709 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hx6gq"] Dec 05 09:23:03 crc kubenswrapper[4815]: I1205 09:23:03.722259 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hx6gq"] Dec 05 09:23:03 crc kubenswrapper[4815]: E1205 09:23:03.830452 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35a4b760_e48f_428b_8ec9_8779a968af92.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod35a4b760_e48f_428b_8ec9_8779a968af92.slice/crio-7a97af43e3f0d05fe20670a742a9de1217241423978b849048a58d9ce4e575d5\": RecentStats: unable to find data in memory cache]" Dec 05 09:23:04 crc kubenswrapper[4815]: I1205 09:23:04.372653 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbdfv" event={"ID":"563ef3a0-d97f-4b5b-afc4-378bdbbee19b","Type":"ContainerStarted","Data":"f8ebea16cb788d77b342556cb2f16eb6a3e0ba5331a4e02a6e2ee6d8bddabc48"} Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.154499 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz"] Dec 05 09:23:05 crc kubenswrapper[4815]: E1205 09:23:05.154990 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" containerName="extract-content" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.155003 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" containerName="extract-content" Dec 05 09:23:05 crc kubenswrapper[4815]: E1205 09:23:05.155014 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35a4b760-e48f-428b-8ec9-8779a968af92" containerName="registry-server" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.155020 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="35a4b760-e48f-428b-8ec9-8779a968af92" containerName="registry-server" Dec 05 09:23:05 crc kubenswrapper[4815]: E1205 09:23:05.155033 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35a4b760-e48f-428b-8ec9-8779a968af92" containerName="extract-content" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.155039 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="35a4b760-e48f-428b-8ec9-8779a968af92" containerName="extract-content" Dec 05 09:23:05 crc kubenswrapper[4815]: E1205 09:23:05.155051 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35a4b760-e48f-428b-8ec9-8779a968af92" containerName="extract-utilities" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.155057 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="35a4b760-e48f-428b-8ec9-8779a968af92" containerName="extract-utilities" Dec 05 09:23:05 crc kubenswrapper[4815]: E1205 09:23:05.155072 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" containerName="registry-server" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.155077 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" containerName="registry-server" Dec 05 09:23:05 crc kubenswrapper[4815]: E1205 09:23:05.155088 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" containerName="extract-utilities" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.155095 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" containerName="extract-utilities" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.155206 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="24cd25c3-6c61-4d4e-8af5-26fefc5405da" containerName="registry-server" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.155216 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="35a4b760-e48f-428b-8ec9-8779a968af92" containerName="registry-server" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.156002 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.158245 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-shkxw" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.167462 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz"] Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.328083 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.328125 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.328165 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zszn8\" (UniqueName: \"kubernetes.io/projected/e115064e-50f8-4b07-962d-fa7e7ea7108f-kube-api-access-zszn8\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.381729 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbdfv" event={"ID":"563ef3a0-d97f-4b5b-afc4-378bdbbee19b","Type":"ContainerStarted","Data":"0e5e6877f3b4ff721b23e81eaff8f10fd0503e931d9812b6058331e048be50a3"} Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.381764 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbdfv" event={"ID":"563ef3a0-d97f-4b5b-afc4-378bdbbee19b","Type":"ContainerStarted","Data":"b3dcaa7472a3373839eb08349e50d910be6285ac26a90ae57b76495288e635be"} Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.381775 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbdfv" event={"ID":"563ef3a0-d97f-4b5b-afc4-378bdbbee19b","Type":"ContainerStarted","Data":"c2f3753bde299293325e6fc191cdf8654014546b121185619971da3e981337b6"} Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.424844 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35a4b760-e48f-428b-8ec9-8779a968af92" path="/var/lib/kubelet/pods/35a4b760-e48f-428b-8ec9-8779a968af92/volumes" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.429739 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.429791 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.429856 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zszn8\" (UniqueName: \"kubernetes.io/projected/e115064e-50f8-4b07-962d-fa7e7ea7108f-kube-api-access-zszn8\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.430184 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-util\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.430456 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-bundle\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.452902 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zszn8\" (UniqueName: \"kubernetes.io/projected/e115064e-50f8-4b07-962d-fa7e7ea7108f-kube-api-access-zszn8\") pod \"917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.471975 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:05 crc kubenswrapper[4815]: I1205 09:23:05.704819 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz"] Dec 05 09:23:05 crc kubenswrapper[4815]: W1205 09:23:05.709114 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode115064e_50f8_4b07_962d_fa7e7ea7108f.slice/crio-dd8e0ba5d2d2e00604f30f4272c1f2f0929281a776040ce0905d49a6a0a15076 WatchSource:0}: Error finding container dd8e0ba5d2d2e00604f30f4272c1f2f0929281a776040ce0905d49a6a0a15076: Status 404 returned error can't find the container with id dd8e0ba5d2d2e00604f30f4272c1f2f0929281a776040ce0905d49a6a0a15076 Dec 05 09:23:06 crc kubenswrapper[4815]: I1205 09:23:06.391172 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbdfv" event={"ID":"563ef3a0-d97f-4b5b-afc4-378bdbbee19b","Type":"ContainerStarted","Data":"c3ba3b20567015fa4fb8fb481b49e956132b6b273fd3b1da80aad928de5153a3"} Dec 05 09:23:06 crc kubenswrapper[4815]: I1205 09:23:06.392249 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" event={"ID":"e115064e-50f8-4b07-962d-fa7e7ea7108f","Type":"ContainerStarted","Data":"dd8e0ba5d2d2e00604f30f4272c1f2f0929281a776040ce0905d49a6a0a15076"} Dec 05 09:23:07 crc kubenswrapper[4815]: I1205 09:23:07.406790 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jbdfv" event={"ID":"563ef3a0-d97f-4b5b-afc4-378bdbbee19b","Type":"ContainerStarted","Data":"121e3c0fc89cda3b462dbd4d0076a92e6ec11d6305879721cb75856fce169b1b"} Dec 05 09:23:07 crc kubenswrapper[4815]: I1205 09:23:07.408826 4815 generic.go:334] "Generic (PLEG): container finished" podID="e115064e-50f8-4b07-962d-fa7e7ea7108f" containerID="3b0dcf7e55a6bc66eeb53e6201ea4bbeb89889d512d529dfe490a3a4c01dddfb" exitCode=0 Dec 05 09:23:07 crc kubenswrapper[4815]: I1205 09:23:07.408859 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" event={"ID":"e115064e-50f8-4b07-962d-fa7e7ea7108f","Type":"ContainerDied","Data":"3b0dcf7e55a6bc66eeb53e6201ea4bbeb89889d512d529dfe490a3a4c01dddfb"} Dec 05 09:23:08 crc kubenswrapper[4815]: I1205 09:23:08.414323 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:23:08 crc kubenswrapper[4815]: I1205 09:23:08.443154 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-jbdfv" podStartSLOduration=-9223371990.411638 podStartE2EDuration="46.443136619s" podCreationTimestamp="2025-12-05 09:22:22 +0000 UTC" firstStartedPulling="2025-12-05 09:22:24.080639116 +0000 UTC m=+942.959245953" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:23:08.439567881 +0000 UTC m=+987.318174728" watchObservedRunningTime="2025-12-05 09:23:08.443136619 +0000 UTC m=+987.321743466" Dec 05 09:23:08 crc kubenswrapper[4815]: I1205 09:23:08.829302 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:23:08 crc kubenswrapper[4815]: I1205 09:23:08.884860 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:23:09 crc kubenswrapper[4815]: I1205 09:23:09.426800 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" event={"ID":"e115064e-50f8-4b07-962d-fa7e7ea7108f","Type":"ContainerStarted","Data":"32a7c509a5bf06925974f05268f8569ca13623e1d175bdab04657b056a0289e2"} Dec 05 09:23:10 crc kubenswrapper[4815]: I1205 09:23:10.432992 4815 generic.go:334] "Generic (PLEG): container finished" podID="e115064e-50f8-4b07-962d-fa7e7ea7108f" containerID="32a7c509a5bf06925974f05268f8569ca13623e1d175bdab04657b056a0289e2" exitCode=0 Dec 05 09:23:10 crc kubenswrapper[4815]: I1205 09:23:10.433655 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" event={"ID":"e115064e-50f8-4b07-962d-fa7e7ea7108f","Type":"ContainerDied","Data":"32a7c509a5bf06925974f05268f8569ca13623e1d175bdab04657b056a0289e2"} Dec 05 09:23:12 crc kubenswrapper[4815]: I1205 09:23:12.446670 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" event={"ID":"e115064e-50f8-4b07-962d-fa7e7ea7108f","Type":"ContainerStarted","Data":"90bc36ee9b92f7c33559c02738ac97ad6c9076dcff811fd9da3952440c189732"} Dec 05 09:23:13 crc kubenswrapper[4815]: I1205 09:23:13.817582 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" Dec 05 09:23:13 crc kubenswrapper[4815]: I1205 09:23:13.837945 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-jbdfv" Dec 05 09:23:17 crc kubenswrapper[4815]: I1205 09:23:17.498768 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" podStartSLOduration=10.720715019 podStartE2EDuration="12.498748389s" podCreationTimestamp="2025-12-05 09:23:05 +0000 UTC" firstStartedPulling="2025-12-05 09:23:07.410024776 +0000 UTC m=+986.288631613" lastFinishedPulling="2025-12-05 09:23:09.188058146 +0000 UTC m=+988.066664983" observedRunningTime="2025-12-05 09:23:17.494641366 +0000 UTC m=+996.373248193" watchObservedRunningTime="2025-12-05 09:23:17.498748389 +0000 UTC m=+996.377355246" Dec 05 09:23:18 crc kubenswrapper[4815]: I1205 09:23:18.484795 4815 generic.go:334] "Generic (PLEG): container finished" podID="e115064e-50f8-4b07-962d-fa7e7ea7108f" containerID="90bc36ee9b92f7c33559c02738ac97ad6c9076dcff811fd9da3952440c189732" exitCode=0 Dec 05 09:23:18 crc kubenswrapper[4815]: I1205 09:23:18.484851 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" event={"ID":"e115064e-50f8-4b07-962d-fa7e7ea7108f","Type":"ContainerDied","Data":"90bc36ee9b92f7c33559c02738ac97ad6c9076dcff811fd9da3952440c189732"} Dec 05 09:23:19 crc kubenswrapper[4815]: I1205 09:23:19.712562 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:19 crc kubenswrapper[4815]: I1205 09:23:19.846362 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-util\") pod \"e115064e-50f8-4b07-962d-fa7e7ea7108f\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " Dec 05 09:23:19 crc kubenswrapper[4815]: I1205 09:23:19.846429 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zszn8\" (UniqueName: \"kubernetes.io/projected/e115064e-50f8-4b07-962d-fa7e7ea7108f-kube-api-access-zszn8\") pod \"e115064e-50f8-4b07-962d-fa7e7ea7108f\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " Dec 05 09:23:19 crc kubenswrapper[4815]: I1205 09:23:19.846506 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-bundle\") pod \"e115064e-50f8-4b07-962d-fa7e7ea7108f\" (UID: \"e115064e-50f8-4b07-962d-fa7e7ea7108f\") " Dec 05 09:23:19 crc kubenswrapper[4815]: I1205 09:23:19.848049 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-bundle" (OuterVolumeSpecName: "bundle") pod "e115064e-50f8-4b07-962d-fa7e7ea7108f" (UID: "e115064e-50f8-4b07-962d-fa7e7ea7108f"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:23:19 crc kubenswrapper[4815]: I1205 09:23:19.851599 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e115064e-50f8-4b07-962d-fa7e7ea7108f-kube-api-access-zszn8" (OuterVolumeSpecName: "kube-api-access-zszn8") pod "e115064e-50f8-4b07-962d-fa7e7ea7108f" (UID: "e115064e-50f8-4b07-962d-fa7e7ea7108f"). InnerVolumeSpecName "kube-api-access-zszn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:23:19 crc kubenswrapper[4815]: I1205 09:23:19.863125 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-util" (OuterVolumeSpecName: "util") pod "e115064e-50f8-4b07-962d-fa7e7ea7108f" (UID: "e115064e-50f8-4b07-962d-fa7e7ea7108f"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:23:19 crc kubenswrapper[4815]: I1205 09:23:19.947560 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zszn8\" (UniqueName: \"kubernetes.io/projected/e115064e-50f8-4b07-962d-fa7e7ea7108f-kube-api-access-zszn8\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:19 crc kubenswrapper[4815]: I1205 09:23:19.947594 4815 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:19 crc kubenswrapper[4815]: I1205 09:23:19.947603 4815 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e115064e-50f8-4b07-962d-fa7e7ea7108f-util\") on node \"crc\" DevicePath \"\"" Dec 05 09:23:20 crc kubenswrapper[4815]: I1205 09:23:20.192950 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:23:20 crc kubenswrapper[4815]: I1205 09:23:20.193013 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:23:20 crc kubenswrapper[4815]: I1205 09:23:20.193066 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:23:20 crc kubenswrapper[4815]: I1205 09:23:20.193720 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bda5cae8332591142b5d1ee8c164d1bfba69b735e07e595565abef36f7e9229e"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:23:20 crc kubenswrapper[4815]: I1205 09:23:20.193790 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://bda5cae8332591142b5d1ee8c164d1bfba69b735e07e595565abef36f7e9229e" gracePeriod=600 Dec 05 09:23:20 crc kubenswrapper[4815]: I1205 09:23:20.497925 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="bda5cae8332591142b5d1ee8c164d1bfba69b735e07e595565abef36f7e9229e" exitCode=0 Dec 05 09:23:20 crc kubenswrapper[4815]: I1205 09:23:20.497978 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"bda5cae8332591142b5d1ee8c164d1bfba69b735e07e595565abef36f7e9229e"} Dec 05 09:23:20 crc kubenswrapper[4815]: I1205 09:23:20.498338 4815 scope.go:117] "RemoveContainer" containerID="0c25c98ba21766bd050ab40ffa71bd173c8e738150c82cbdfb30f46550e857a7" Dec 05 09:23:20 crc kubenswrapper[4815]: I1205 09:23:20.500851 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" event={"ID":"e115064e-50f8-4b07-962d-fa7e7ea7108f","Type":"ContainerDied","Data":"dd8e0ba5d2d2e00604f30f4272c1f2f0929281a776040ce0905d49a6a0a15076"} Dec 05 09:23:20 crc kubenswrapper[4815]: I1205 09:23:20.500975 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd8e0ba5d2d2e00604f30f4272c1f2f0929281a776040ce0905d49a6a0a15076" Dec 05 09:23:20 crc kubenswrapper[4815]: I1205 09:23:20.500919 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz" Dec 05 09:23:22 crc kubenswrapper[4815]: I1205 09:23:22.519895 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"e422e582a96b96c6d83e3c0fc7b930a3d237267c29a704a2f4827be174a6d09c"} Dec 05 09:23:27 crc kubenswrapper[4815]: I1205 09:23:27.829128 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd"] Dec 05 09:23:27 crc kubenswrapper[4815]: E1205 09:23:27.829691 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e115064e-50f8-4b07-962d-fa7e7ea7108f" containerName="pull" Dec 05 09:23:27 crc kubenswrapper[4815]: I1205 09:23:27.829712 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e115064e-50f8-4b07-962d-fa7e7ea7108f" containerName="pull" Dec 05 09:23:27 crc kubenswrapper[4815]: E1205 09:23:27.829737 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e115064e-50f8-4b07-962d-fa7e7ea7108f" containerName="extract" Dec 05 09:23:27 crc kubenswrapper[4815]: I1205 09:23:27.829745 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e115064e-50f8-4b07-962d-fa7e7ea7108f" containerName="extract" Dec 05 09:23:27 crc kubenswrapper[4815]: E1205 09:23:27.829760 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e115064e-50f8-4b07-962d-fa7e7ea7108f" containerName="util" Dec 05 09:23:27 crc kubenswrapper[4815]: I1205 09:23:27.829767 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e115064e-50f8-4b07-962d-fa7e7ea7108f" containerName="util" Dec 05 09:23:27 crc kubenswrapper[4815]: I1205 09:23:27.829914 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e115064e-50f8-4b07-962d-fa7e7ea7108f" containerName="extract" Dec 05 09:23:27 crc kubenswrapper[4815]: I1205 09:23:27.830442 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd" Dec 05 09:23:27 crc kubenswrapper[4815]: I1205 09:23:27.844054 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-cp74r" Dec 05 09:23:27 crc kubenswrapper[4815]: I1205 09:23:27.907910 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd"] Dec 05 09:23:27 crc kubenswrapper[4815]: I1205 09:23:27.972822 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z55p\" (UniqueName: \"kubernetes.io/projected/7f520180-b5ab-420b-9479-4089e1845ecc-kube-api-access-4z55p\") pod \"openstack-operator-controller-operator-55b6fb9447-xg6vd\" (UID: \"7f520180-b5ab-420b-9479-4089e1845ecc\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd" Dec 05 09:23:28 crc kubenswrapper[4815]: I1205 09:23:28.074102 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z55p\" (UniqueName: \"kubernetes.io/projected/7f520180-b5ab-420b-9479-4089e1845ecc-kube-api-access-4z55p\") pod \"openstack-operator-controller-operator-55b6fb9447-xg6vd\" (UID: \"7f520180-b5ab-420b-9479-4089e1845ecc\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd" Dec 05 09:23:28 crc kubenswrapper[4815]: I1205 09:23:28.148984 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z55p\" (UniqueName: \"kubernetes.io/projected/7f520180-b5ab-420b-9479-4089e1845ecc-kube-api-access-4z55p\") pod \"openstack-operator-controller-operator-55b6fb9447-xg6vd\" (UID: \"7f520180-b5ab-420b-9479-4089e1845ecc\") " pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd" Dec 05 09:23:28 crc kubenswrapper[4815]: I1205 09:23:28.449290 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd" Dec 05 09:23:28 crc kubenswrapper[4815]: I1205 09:23:28.771973 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd"] Dec 05 09:23:28 crc kubenswrapper[4815]: W1205 09:23:28.773861 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f520180_b5ab_420b_9479_4089e1845ecc.slice/crio-d3e3368438824922bdde9c82fa50cefec76dd8776d8d85f8d50aa658ce008735 WatchSource:0}: Error finding container d3e3368438824922bdde9c82fa50cefec76dd8776d8d85f8d50aa658ce008735: Status 404 returned error can't find the container with id d3e3368438824922bdde9c82fa50cefec76dd8776d8d85f8d50aa658ce008735 Dec 05 09:23:28 crc kubenswrapper[4815]: I1205 09:23:28.776892 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:23:29 crc kubenswrapper[4815]: I1205 09:23:29.607286 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd" event={"ID":"7f520180-b5ab-420b-9479-4089e1845ecc","Type":"ContainerStarted","Data":"d3e3368438824922bdde9c82fa50cefec76dd8776d8d85f8d50aa658ce008735"} Dec 05 09:23:39 crc kubenswrapper[4815]: I1205 09:23:39.685511 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd" event={"ID":"7f520180-b5ab-420b-9479-4089e1845ecc","Type":"ContainerStarted","Data":"12c250a4a67982a769a19466247ae10b08739f9b3c48628862d10ba1c6aa7f78"} Dec 05 09:23:39 crc kubenswrapper[4815]: I1205 09:23:39.686133 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd" Dec 05 09:23:39 crc kubenswrapper[4815]: I1205 09:23:39.728936 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd" podStartSLOduration=2.085583244 podStartE2EDuration="12.72891804s" podCreationTimestamp="2025-12-05 09:23:27 +0000 UTC" firstStartedPulling="2025-12-05 09:23:28.776607158 +0000 UTC m=+1007.655213995" lastFinishedPulling="2025-12-05 09:23:39.419941934 +0000 UTC m=+1018.298548791" observedRunningTime="2025-12-05 09:23:39.722673239 +0000 UTC m=+1018.601280076" watchObservedRunningTime="2025-12-05 09:23:39.72891804 +0000 UTC m=+1018.607524877" Dec 05 09:23:48 crc kubenswrapper[4815]: I1205 09:23:48.452044 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-55b6fb9447-xg6vd" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.652042 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.654579 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.657082 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-hsc2g" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.663937 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.665005 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.668709 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.673055 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-qpt77" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.688383 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.731664 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.732969 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.736719 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-krc2x" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.748181 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.749461 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.756180 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-lb6km" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.766817 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72vrb\" (UniqueName: \"kubernetes.io/projected/cafb574b-a659-41e6-89d5-c02851274f73-kube-api-access-72vrb\") pod \"barbican-operator-controller-manager-7d9dfd778-26tt9\" (UID: \"cafb574b-a659-41e6-89d5-c02851274f73\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.766865 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzr9m\" (UniqueName: \"kubernetes.io/projected/f023493d-1716-40ac-816d-26067f0019f3-kube-api-access-bzr9m\") pod \"cinder-operator-controller-manager-859b6ccc6-j8cxc\" (UID: \"f023493d-1716-40ac-816d-26067f0019f3\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.778168 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.782797 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.798597 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.799984 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.810049 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-t6jb6" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.830699 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.852639 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.853887 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.859107 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-4tdrv" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.870337 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx9xt\" (UniqueName: \"kubernetes.io/projected/f23f3b0d-7874-4906-81de-b8d9226082d2-kube-api-access-zx9xt\") pod \"heat-operator-controller-manager-5f64f6f8bb-tmg9c\" (UID: \"f23f3b0d-7874-4906-81de-b8d9226082d2\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.870383 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72vrb\" (UniqueName: \"kubernetes.io/projected/cafb574b-a659-41e6-89d5-c02851274f73-kube-api-access-72vrb\") pod \"barbican-operator-controller-manager-7d9dfd778-26tt9\" (UID: \"cafb574b-a659-41e6-89d5-c02851274f73\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.870415 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cz2x\" (UniqueName: \"kubernetes.io/projected/9db099f9-ed44-46ed-ab3d-97d09ee0b5b0-kube-api-access-6cz2x\") pod \"designate-operator-controller-manager-78b4bc895b-g4nbp\" (UID: \"9db099f9-ed44-46ed-ab3d-97d09ee0b5b0\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.870440 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzr9m\" (UniqueName: \"kubernetes.io/projected/f023493d-1716-40ac-816d-26067f0019f3-kube-api-access-bzr9m\") pod \"cinder-operator-controller-manager-859b6ccc6-j8cxc\" (UID: \"f023493d-1716-40ac-816d-26067f0019f3\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.870468 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq7pc\" (UniqueName: \"kubernetes.io/projected/591bafca-5706-418f-b803-d29d1b0865db-kube-api-access-vq7pc\") pod \"glance-operator-controller-manager-77987cd8cd-8blh4\" (UID: \"591bafca-5706-418f-b803-d29d1b0865db\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.884033 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.912332 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-9pflh"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.913915 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.920829 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzr9m\" (UniqueName: \"kubernetes.io/projected/f023493d-1716-40ac-816d-26067f0019f3-kube-api-access-bzr9m\") pod \"cinder-operator-controller-manager-859b6ccc6-j8cxc\" (UID: \"f023493d-1716-40ac-816d-26067f0019f3\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.925696 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.933043 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.936175 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.937317 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-vdsgk" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.937472 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-l4l8f" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.940291 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72vrb\" (UniqueName: \"kubernetes.io/projected/cafb574b-a659-41e6-89d5-c02851274f73-kube-api-access-72vrb\") pod \"barbican-operator-controller-manager-7d9dfd778-26tt9\" (UID: \"cafb574b-a659-41e6-89d5-c02851274f73\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.956844 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-9pflh"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.962369 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff"] Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.974002 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbf6z\" (UniqueName: \"kubernetes.io/projected/ed463fa2-6734-41b5-a34b-cb83436130fd-kube-api-access-fbf6z\") pod \"horizon-operator-controller-manager-68c6d99b8f-5jh5q\" (UID: \"ed463fa2-6734-41b5-a34b-cb83436130fd\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.974073 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert\") pod \"infra-operator-controller-manager-57548d458d-9pflh\" (UID: \"4abce90f-180f-4302-aecd-89bf5739757c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.974119 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx9xt\" (UniqueName: \"kubernetes.io/projected/f23f3b0d-7874-4906-81de-b8d9226082d2-kube-api-access-zx9xt\") pod \"heat-operator-controller-manager-5f64f6f8bb-tmg9c\" (UID: \"f23f3b0d-7874-4906-81de-b8d9226082d2\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.974150 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cz2x\" (UniqueName: \"kubernetes.io/projected/9db099f9-ed44-46ed-ab3d-97d09ee0b5b0-kube-api-access-6cz2x\") pod \"designate-operator-controller-manager-78b4bc895b-g4nbp\" (UID: \"9db099f9-ed44-46ed-ab3d-97d09ee0b5b0\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.974203 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq7pc\" (UniqueName: \"kubernetes.io/projected/591bafca-5706-418f-b803-d29d1b0865db-kube-api-access-vq7pc\") pod \"glance-operator-controller-manager-77987cd8cd-8blh4\" (UID: \"591bafca-5706-418f-b803-d29d1b0865db\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.974243 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zvg5\" (UniqueName: \"kubernetes.io/projected/4abce90f-180f-4302-aecd-89bf5739757c-kube-api-access-7zvg5\") pod \"infra-operator-controller-manager-57548d458d-9pflh\" (UID: \"4abce90f-180f-4302-aecd-89bf5739757c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:07 crc kubenswrapper[4815]: I1205 09:24:07.974959 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.001862 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.027643 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx9xt\" (UniqueName: \"kubernetes.io/projected/f23f3b0d-7874-4906-81de-b8d9226082d2-kube-api-access-zx9xt\") pod \"heat-operator-controller-manager-5f64f6f8bb-tmg9c\" (UID: \"f23f3b0d-7874-4906-81de-b8d9226082d2\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.148313 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.150447 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq7pc\" (UniqueName: \"kubernetes.io/projected/591bafca-5706-418f-b803-d29d1b0865db-kube-api-access-vq7pc\") pod \"glance-operator-controller-manager-77987cd8cd-8blh4\" (UID: \"591bafca-5706-418f-b803-d29d1b0865db\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.153815 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zvg5\" (UniqueName: \"kubernetes.io/projected/4abce90f-180f-4302-aecd-89bf5739757c-kube-api-access-7zvg5\") pod \"infra-operator-controller-manager-57548d458d-9pflh\" (UID: \"4abce90f-180f-4302-aecd-89bf5739757c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.153957 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9kdm\" (UniqueName: \"kubernetes.io/projected/ef1e93b5-9a31-4588-a5e4-78dbe6ef0437-kube-api-access-g9kdm\") pod \"ironic-operator-controller-manager-6c548fd776-7mtff\" (UID: \"ef1e93b5-9a31-4588-a5e4-78dbe6ef0437\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.154014 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbf6z\" (UniqueName: \"kubernetes.io/projected/ed463fa2-6734-41b5-a34b-cb83436130fd-kube-api-access-fbf6z\") pod \"horizon-operator-controller-manager-68c6d99b8f-5jh5q\" (UID: \"ed463fa2-6734-41b5-a34b-cb83436130fd\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.154168 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert\") pod \"infra-operator-controller-manager-57548d458d-9pflh\" (UID: \"4abce90f-180f-4302-aecd-89bf5739757c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:08 crc kubenswrapper[4815]: E1205 09:24:08.154904 4815 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 09:24:08 crc kubenswrapper[4815]: E1205 09:24:08.155086 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert podName:4abce90f-180f-4302-aecd-89bf5739757c nodeName:}" failed. No retries permitted until 2025-12-05 09:24:08.654945596 +0000 UTC m=+1047.533552433 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert") pod "infra-operator-controller-manager-57548d458d-9pflh" (UID: "4abce90f-180f-4302-aecd-89bf5739757c") : secret "infra-operator-webhook-server-cert" not found Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.157585 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cz2x\" (UniqueName: \"kubernetes.io/projected/9db099f9-ed44-46ed-ab3d-97d09ee0b5b0-kube-api-access-6cz2x\") pod \"designate-operator-controller-manager-78b4bc895b-g4nbp\" (UID: \"9db099f9-ed44-46ed-ab3d-97d09ee0b5b0\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.232054 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.233982 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.244317 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.245360 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.253764 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-4nmq6" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.256648 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-f84rs" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.259469 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.266092 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9kdm\" (UniqueName: \"kubernetes.io/projected/ef1e93b5-9a31-4588-a5e4-78dbe6ef0437-kube-api-access-g9kdm\") pod \"ironic-operator-controller-manager-6c548fd776-7mtff\" (UID: \"ef1e93b5-9a31-4588-a5e4-78dbe6ef0437\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.268021 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.268851 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.269111 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.274556 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.276533 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zvg5\" (UniqueName: \"kubernetes.io/projected/4abce90f-180f-4302-aecd-89bf5739757c-kube-api-access-7zvg5\") pod \"infra-operator-controller-manager-57548d458d-9pflh\" (UID: \"4abce90f-180f-4302-aecd-89bf5739757c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.277294 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbf6z\" (UniqueName: \"kubernetes.io/projected/ed463fa2-6734-41b5-a34b-cb83436130fd-kube-api-access-fbf6z\") pod \"horizon-operator-controller-manager-68c6d99b8f-5jh5q\" (UID: \"ed463fa2-6734-41b5-a34b-cb83436130fd\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.300423 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-c9g5b" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.304571 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-ddr46" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.307193 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.327341 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.347803 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9kdm\" (UniqueName: \"kubernetes.io/projected/ef1e93b5-9a31-4588-a5e4-78dbe6ef0437-kube-api-access-g9kdm\") pod \"ironic-operator-controller-manager-6c548fd776-7mtff\" (UID: \"ef1e93b5-9a31-4588-a5e4-78dbe6ef0437\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.351260 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.352226 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.368031 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.368398 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwzzd\" (UniqueName: \"kubernetes.io/projected/cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb-kube-api-access-zwzzd\") pod \"keystone-operator-controller-manager-7765d96ddf-w7vjm\" (UID: \"cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.368912 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-45f2d" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.371272 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.376261 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwz4g\" (UniqueName: \"kubernetes.io/projected/f80008bd-769f-4960-be8f-62894fdd7718-kube-api-access-kwz4g\") pod \"manila-operator-controller-manager-7c79b5df47-2k8wt\" (UID: \"f80008bd-769f-4960-be8f-62894fdd7718\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.376305 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnxrc\" (UniqueName: \"kubernetes.io/projected/7482e78f-1387-4e91-b5d4-419f5164fea5-kube-api-access-cnxrc\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-ccd95\" (UID: \"7482e78f-1387-4e91-b5d4-419f5164fea5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.376351 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9xrp\" (UniqueName: \"kubernetes.io/projected/3665adde-25ad-4261-be50-ef4e21726b7b-kube-api-access-k9xrp\") pod \"mariadb-operator-controller-manager-56bbcc9d85-bm8kq\" (UID: \"3665adde-25ad-4261-be50-ef4e21726b7b\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.386699 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.394092 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.395418 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.421961 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.436878 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-d9mtc" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.469628 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.470740 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.471538 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.481885 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgf4s\" (UniqueName: \"kubernetes.io/projected/48f9720a-5d73-4773-85f9-9c571c9322ee-kube-api-access-pgf4s\") pod \"nova-operator-controller-manager-697bc559fc-25lnj\" (UID: \"48f9720a-5d73-4773-85f9-9c571c9322ee\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.481944 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r72wl\" (UniqueName: \"kubernetes.io/projected/b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d-kube-api-access-r72wl\") pod \"octavia-operator-controller-manager-998648c74-l5qqd\" (UID: \"b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.481985 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwz4g\" (UniqueName: \"kubernetes.io/projected/f80008bd-769f-4960-be8f-62894fdd7718-kube-api-access-kwz4g\") pod \"manila-operator-controller-manager-7c79b5df47-2k8wt\" (UID: \"f80008bd-769f-4960-be8f-62894fdd7718\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.482003 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnxrc\" (UniqueName: \"kubernetes.io/projected/7482e78f-1387-4e91-b5d4-419f5164fea5-kube-api-access-cnxrc\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-ccd95\" (UID: \"7482e78f-1387-4e91-b5d4-419f5164fea5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.482027 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9xrp\" (UniqueName: \"kubernetes.io/projected/3665adde-25ad-4261-be50-ef4e21726b7b-kube-api-access-k9xrp\") pod \"mariadb-operator-controller-manager-56bbcc9d85-bm8kq\" (UID: \"3665adde-25ad-4261-be50-ef4e21726b7b\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.482085 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwzzd\" (UniqueName: \"kubernetes.io/projected/cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb-kube-api-access-zwzzd\") pod \"keystone-operator-controller-manager-7765d96ddf-w7vjm\" (UID: \"cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.495865 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-gf42k" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.534721 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.536116 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.540789 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-r7bkd" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.545981 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.558061 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.559039 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.566541 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwzzd\" (UniqueName: \"kubernetes.io/projected/cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb-kube-api-access-zwzzd\") pod \"keystone-operator-controller-manager-7765d96ddf-w7vjm\" (UID: \"cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.587788 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.588790 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.589596 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgf4s\" (UniqueName: \"kubernetes.io/projected/48f9720a-5d73-4773-85f9-9c571c9322ee-kube-api-access-pgf4s\") pod \"nova-operator-controller-manager-697bc559fc-25lnj\" (UID: \"48f9720a-5d73-4773-85f9-9c571c9322ee\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.589623 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgwvm\" (UniqueName: \"kubernetes.io/projected/378923e7-18f6-4946-8e86-d1ececca15cc-kube-api-access-jgwvm\") pod \"ovn-operator-controller-manager-b6456fdb6-5jlqr\" (UID: \"378923e7-18f6-4946-8e86-d1ececca15cc\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.589652 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r72wl\" (UniqueName: \"kubernetes.io/projected/b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d-kube-api-access-r72wl\") pod \"octavia-operator-controller-manager-998648c74-l5qqd\" (UID: \"b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.603728 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-tzh6z" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.605759 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.615813 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.618949 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-wdf2r" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.622401 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.631362 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnxrc\" (UniqueName: \"kubernetes.io/projected/7482e78f-1387-4e91-b5d4-419f5164fea5-kube-api-access-cnxrc\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-ccd95\" (UID: \"7482e78f-1387-4e91-b5d4-419f5164fea5\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.643057 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9xrp\" (UniqueName: \"kubernetes.io/projected/3665adde-25ad-4261-be50-ef4e21726b7b-kube-api-access-k9xrp\") pod \"mariadb-operator-controller-manager-56bbcc9d85-bm8kq\" (UID: \"3665adde-25ad-4261-be50-ef4e21726b7b\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.651501 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r72wl\" (UniqueName: \"kubernetes.io/projected/b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d-kube-api-access-r72wl\") pod \"octavia-operator-controller-manager-998648c74-l5qqd\" (UID: \"b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.653153 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwz4g\" (UniqueName: \"kubernetes.io/projected/f80008bd-769f-4960-be8f-62894fdd7718-kube-api-access-kwz4g\") pod \"manila-operator-controller-manager-7c79b5df47-2k8wt\" (UID: \"f80008bd-769f-4960-be8f-62894fdd7718\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.659849 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.678111 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgf4s\" (UniqueName: \"kubernetes.io/projected/48f9720a-5d73-4773-85f9-9c571c9322ee-kube-api-access-pgf4s\") pod \"nova-operator-controller-manager-697bc559fc-25lnj\" (UID: \"48f9720a-5d73-4773-85f9-9c571c9322ee\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.692585 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.693355 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbh5t\" (UniqueName: \"kubernetes.io/projected/b7558b70-107c-48ea-ac22-e42b1bcdf47d-kube-api-access-hbh5t\") pod \"openstack-baremetal-operator-controller-manager-55c85496f5b4xkh\" (UID: \"b7558b70-107c-48ea-ac22-e42b1bcdf47d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.693469 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f5b4xkh\" (UID: \"b7558b70-107c-48ea-ac22-e42b1bcdf47d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.693586 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.693681 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mck75\" (UniqueName: \"kubernetes.io/projected/47664970-bfba-453a-ad06-ac1e03417eac-kube-api-access-mck75\") pod \"swift-operator-controller-manager-5f8c65bbfc-6nl6f\" (UID: \"47664970-bfba-453a-ad06-ac1e03417eac\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.693770 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert\") pod \"infra-operator-controller-manager-57548d458d-9pflh\" (UID: \"4abce90f-180f-4302-aecd-89bf5739757c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.693879 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgwvm\" (UniqueName: \"kubernetes.io/projected/378923e7-18f6-4946-8e86-d1ececca15cc-kube-api-access-jgwvm\") pod \"ovn-operator-controller-manager-b6456fdb6-5jlqr\" (UID: \"378923e7-18f6-4946-8e86-d1ececca15cc\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.693986 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v28xz\" (UniqueName: \"kubernetes.io/projected/6b41117c-2922-4337-b00c-2bebd57a2c6b-kube-api-access-v28xz\") pod \"placement-operator-controller-manager-78f8948974-8pt5c\" (UID: \"6b41117c-2922-4337-b00c-2bebd57a2c6b\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" Dec 05 09:24:08 crc kubenswrapper[4815]: E1205 09:24:08.693989 4815 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 09:24:08 crc kubenswrapper[4815]: E1205 09:24:08.694185 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert podName:4abce90f-180f-4302-aecd-89bf5739757c nodeName:}" failed. No retries permitted until 2025-12-05 09:24:09.694165103 +0000 UTC m=+1048.572771940 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert") pod "infra-operator-controller-manager-57548d458d-9pflh" (UID: "4abce90f-180f-4302-aecd-89bf5739757c") : secret "infra-operator-webhook-server-cert" not found Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.704691 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.705054 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-hswfr" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.725156 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.736912 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.738750 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.757244 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.774511 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.795983 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v28xz\" (UniqueName: \"kubernetes.io/projected/6b41117c-2922-4337-b00c-2bebd57a2c6b-kube-api-access-v28xz\") pod \"placement-operator-controller-manager-78f8948974-8pt5c\" (UID: \"6b41117c-2922-4337-b00c-2bebd57a2c6b\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.796035 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbh5t\" (UniqueName: \"kubernetes.io/projected/b7558b70-107c-48ea-ac22-e42b1bcdf47d-kube-api-access-hbh5t\") pod \"openstack-baremetal-operator-controller-manager-55c85496f5b4xkh\" (UID: \"b7558b70-107c-48ea-ac22-e42b1bcdf47d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.796063 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f5b4xkh\" (UID: \"b7558b70-107c-48ea-ac22-e42b1bcdf47d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.796104 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mck75\" (UniqueName: \"kubernetes.io/projected/47664970-bfba-453a-ad06-ac1e03417eac-kube-api-access-mck75\") pod \"swift-operator-controller-manager-5f8c65bbfc-6nl6f\" (UID: \"47664970-bfba-453a-ad06-ac1e03417eac\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.796156 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5hgl\" (UniqueName: \"kubernetes.io/projected/926f3f91-cee1-42ae-888f-73ec7f20dd61-kube-api-access-z5hgl\") pod \"telemetry-operator-controller-manager-76cc84c6bb-bnb4c\" (UID: \"926f3f91-cee1-42ae-888f-73ec7f20dd61\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" Dec 05 09:24:08 crc kubenswrapper[4815]: E1205 09:24:08.796689 4815 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 09:24:08 crc kubenswrapper[4815]: E1205 09:24:08.796727 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert podName:b7558b70-107c-48ea-ac22-e42b1bcdf47d nodeName:}" failed. No retries permitted until 2025-12-05 09:24:09.296712992 +0000 UTC m=+1048.175319829 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" (UID: "b7558b70-107c-48ea-ac22-e42b1bcdf47d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.797211 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.808252 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgwvm\" (UniqueName: \"kubernetes.io/projected/378923e7-18f6-4946-8e86-d1ececca15cc-kube-api-access-jgwvm\") pod \"ovn-operator-controller-manager-b6456fdb6-5jlqr\" (UID: \"378923e7-18f6-4946-8e86-d1ececca15cc\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.820591 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.832716 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.846910 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mck75\" (UniqueName: \"kubernetes.io/projected/47664970-bfba-453a-ad06-ac1e03417eac-kube-api-access-mck75\") pod \"swift-operator-controller-manager-5f8c65bbfc-6nl6f\" (UID: \"47664970-bfba-453a-ad06-ac1e03417eac\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.853985 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbh5t\" (UniqueName: \"kubernetes.io/projected/b7558b70-107c-48ea-ac22-e42b1bcdf47d-kube-api-access-hbh5t\") pod \"openstack-baremetal-operator-controller-manager-55c85496f5b4xkh\" (UID: \"b7558b70-107c-48ea-ac22-e42b1bcdf47d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.865186 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v28xz\" (UniqueName: \"kubernetes.io/projected/6b41117c-2922-4337-b00c-2bebd57a2c6b-kube-api-access-v28xz\") pod \"placement-operator-controller-manager-78f8948974-8pt5c\" (UID: \"6b41117c-2922-4337-b00c-2bebd57a2c6b\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.878258 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.888235 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.889232 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.904167 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5hgl\" (UniqueName: \"kubernetes.io/projected/926f3f91-cee1-42ae-888f-73ec7f20dd61-kube-api-access-z5hgl\") pod \"telemetry-operator-controller-manager-76cc84c6bb-bnb4c\" (UID: \"926f3f91-cee1-42ae-888f-73ec7f20dd61\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.904722 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-h46qb" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.905634 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.906867 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.909289 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-l2p5h" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.949954 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.963867 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f"] Dec 05 09:24:08 crc kubenswrapper[4815]: I1205 09:24:08.964397 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.035779 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-94vpt\" (UniqueName: \"kubernetes.io/projected/d8b41c8b-030e-4082-98a3-a0e9fc05a208-kube-api-access-94vpt\") pod \"test-operator-controller-manager-5854674fcc-gbkbf\" (UID: \"d8b41c8b-030e-4082-98a3-a0e9fc05a208\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.035886 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w96l8\" (UniqueName: \"kubernetes.io/projected/19f20867-5315-4be3-836b-57bb0b501b36-kube-api-access-w96l8\") pod \"watcher-operator-controller-manager-769dc69bc-j4r7f\" (UID: \"19f20867-5315-4be3-836b-57bb0b501b36\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.035932 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5hgl\" (UniqueName: \"kubernetes.io/projected/926f3f91-cee1-42ae-888f-73ec7f20dd61-kube-api-access-z5hgl\") pod \"telemetry-operator-controller-manager-76cc84c6bb-bnb4c\" (UID: \"926f3f91-cee1-42ae-888f-73ec7f20dd61\") " pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.055733 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.100296 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf"] Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.153111 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w96l8\" (UniqueName: \"kubernetes.io/projected/19f20867-5315-4be3-836b-57bb0b501b36-kube-api-access-w96l8\") pod \"watcher-operator-controller-manager-769dc69bc-j4r7f\" (UID: \"19f20867-5315-4be3-836b-57bb0b501b36\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.153404 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-94vpt\" (UniqueName: \"kubernetes.io/projected/d8b41c8b-030e-4082-98a3-a0e9fc05a208-kube-api-access-94vpt\") pod \"test-operator-controller-manager-5854674fcc-gbkbf\" (UID: \"d8b41c8b-030e-4082-98a3-a0e9fc05a208\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.186844 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w96l8\" (UniqueName: \"kubernetes.io/projected/19f20867-5315-4be3-836b-57bb0b501b36-kube-api-access-w96l8\") pod \"watcher-operator-controller-manager-769dc69bc-j4r7f\" (UID: \"19f20867-5315-4be3-836b-57bb0b501b36\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.231081 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-94vpt\" (UniqueName: \"kubernetes.io/projected/d8b41c8b-030e-4082-98a3-a0e9fc05a208-kube-api-access-94vpt\") pod \"test-operator-controller-manager-5854674fcc-gbkbf\" (UID: \"d8b41c8b-030e-4082-98a3-a0e9fc05a208\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.260316 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9"] Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.306643 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm"] Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.307708 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.312971 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.316391 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.316602 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-r9dpp" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.338831 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm"] Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.344139 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.356898 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.357337 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f5b4xkh\" (UID: \"b7558b70-107c-48ea-ac22-e42b1bcdf47d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.357426 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqxsv\" (UniqueName: \"kubernetes.io/projected/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-kube-api-access-xqxsv\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.357510 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.357531 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.357712 4815 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.357769 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert podName:b7558b70-107c-48ea-ac22-e42b1bcdf47d nodeName:}" failed. No retries permitted until 2025-12-05 09:24:10.357756008 +0000 UTC m=+1049.236362845 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" (UID: "b7558b70-107c-48ea-ac22-e42b1bcdf47d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.413259 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws"] Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.414201 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.419812 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-pjnwp" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.459349 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.459386 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.459441 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2s7rs\" (UniqueName: \"kubernetes.io/projected/888d282d-8aa1-4a36-9286-d00af63950ba-kube-api-access-2s7rs\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mvkws\" (UID: \"888d282d-8aa1-4a36-9286-d00af63950ba\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.459527 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqxsv\" (UniqueName: \"kubernetes.io/projected/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-kube-api-access-xqxsv\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.460439 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.460521 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs podName:8f0cbf2b-67d6-4058-b173-d24117dfcdd7 nodeName:}" failed. No retries permitted until 2025-12-05 09:24:09.960499731 +0000 UTC m=+1048.839106568 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-2lknm" (UID: "8f0cbf2b-67d6-4058-b173-d24117dfcdd7") : secret "webhook-server-cert" not found Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.460579 4815 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.460606 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs podName:8f0cbf2b-67d6-4058-b173-d24117dfcdd7 nodeName:}" failed. No retries permitted until 2025-12-05 09:24:09.960598334 +0000 UTC m=+1048.839205251 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-2lknm" (UID: "8f0cbf2b-67d6-4058-b173-d24117dfcdd7") : secret "metrics-server-cert" not found Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.491008 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws"] Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.491049 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc"] Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.520212 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqxsv\" (UniqueName: \"kubernetes.io/projected/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-kube-api-access-xqxsv\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.561906 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2s7rs\" (UniqueName: \"kubernetes.io/projected/888d282d-8aa1-4a36-9286-d00af63950ba-kube-api-access-2s7rs\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mvkws\" (UID: \"888d282d-8aa1-4a36-9286-d00af63950ba\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.582135 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q"] Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.586130 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2s7rs\" (UniqueName: \"kubernetes.io/projected/888d282d-8aa1-4a36-9286-d00af63950ba-kube-api-access-2s7rs\") pod \"rabbitmq-cluster-operator-manager-668c99d594-mvkws\" (UID: \"888d282d-8aa1-4a36-9286-d00af63950ba\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.768962 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert\") pod \"infra-operator-controller-manager-57548d458d-9pflh\" (UID: \"4abce90f-180f-4302-aecd-89bf5739757c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.769137 4815 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.769186 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert podName:4abce90f-180f-4302-aecd-89bf5739757c nodeName:}" failed. No retries permitted until 2025-12-05 09:24:11.769171593 +0000 UTC m=+1050.647778430 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert") pod "infra-operator-controller-manager-57548d458d-9pflh" (UID: "4abce90f-180f-4302-aecd-89bf5739757c") : secret "infra-operator-webhook-server-cert" not found Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.769228 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.816103 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4"] Dec 05 09:24:09 crc kubenswrapper[4815]: W1205 09:24:09.853792 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod591bafca_5706_418f_b803_d29d1b0865db.slice/crio-dd5f0000bfcf06cc6fc38648ca9953f1728536c9a1a36ac4c28b43138a1ce92e WatchSource:0}: Error finding container dd5f0000bfcf06cc6fc38648ca9953f1728536c9a1a36ac4c28b43138a1ce92e: Status 404 returned error can't find the container with id dd5f0000bfcf06cc6fc38648ca9953f1728536c9a1a36ac4c28b43138a1ce92e Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.973226 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:09 crc kubenswrapper[4815]: I1205 09:24:09.973268 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.973409 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.973472 4815 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.973476 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs podName:8f0cbf2b-67d6-4058-b173-d24117dfcdd7 nodeName:}" failed. No retries permitted until 2025-12-05 09:24:10.973458237 +0000 UTC m=+1049.852065074 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-2lknm" (UID: "8f0cbf2b-67d6-4058-b173-d24117dfcdd7") : secret "webhook-server-cert" not found Dec 05 09:24:09 crc kubenswrapper[4815]: E1205 09:24:09.973553 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs podName:8f0cbf2b-67d6-4058-b173-d24117dfcdd7 nodeName:}" failed. No retries permitted until 2025-12-05 09:24:10.973535289 +0000 UTC m=+1049.852142126 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-2lknm" (UID: "8f0cbf2b-67d6-4058-b173-d24117dfcdd7") : secret "metrics-server-cert" not found Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.098265 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" event={"ID":"ed463fa2-6734-41b5-a34b-cb83436130fd","Type":"ContainerStarted","Data":"8861c53e2972d1758d507b04ee981f20d5bfaf7bebcd5c7872e3bfc5bd7fc6b5"} Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.113177 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" event={"ID":"cafb574b-a659-41e6-89d5-c02851274f73","Type":"ContainerStarted","Data":"02f0beacf050f51e4e87f956fb757eb7fb78db8f71615e64c4df0c300e756b69"} Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.114866 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" event={"ID":"f023493d-1716-40ac-816d-26067f0019f3","Type":"ContainerStarted","Data":"e79834d6a0abd27a6b708eed4a0abe5248a3556f4d9101a1341130de2dc3e38e"} Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.116200 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" event={"ID":"591bafca-5706-418f-b803-d29d1b0865db","Type":"ContainerStarted","Data":"dd5f0000bfcf06cc6fc38648ca9953f1728536c9a1a36ac4c28b43138a1ce92e"} Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.130197 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c"] Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.359664 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff"] Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.386350 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm"] Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.396980 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f5b4xkh\" (UID: \"b7558b70-107c-48ea-ac22-e42b1bcdf47d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:10 crc kubenswrapper[4815]: E1205 09:24:10.397189 4815 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 09:24:10 crc kubenswrapper[4815]: E1205 09:24:10.397255 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert podName:b7558b70-107c-48ea-ac22-e42b1bcdf47d nodeName:}" failed. No retries permitted until 2025-12-05 09:24:12.397235732 +0000 UTC m=+1051.275842569 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" (UID: "b7558b70-107c-48ea-ac22-e42b1bcdf47d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.403313 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt"] Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.480373 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd"] Dec 05 09:24:10 crc kubenswrapper[4815]: W1205 09:24:10.480575 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3665adde_25ad_4261_be50_ef4e21726b7b.slice/crio-9e7fe6ab5088b598e432997cee09ae1e54137d82e1c30a37fb6afb7e6e9a6cdf WatchSource:0}: Error finding container 9e7fe6ab5088b598e432997cee09ae1e54137d82e1c30a37fb6afb7e6e9a6cdf: Status 404 returned error can't find the container with id 9e7fe6ab5088b598e432997cee09ae1e54137d82e1c30a37fb6afb7e6e9a6cdf Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.487386 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq"] Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.511762 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr"] Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.528270 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj"] Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.559698 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95"] Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.567997 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c"] Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.676663 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c"] Dec 05 09:24:10 crc kubenswrapper[4815]: W1205 09:24:10.695768 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b41117c_2922_4337_b00c_2bebd57a2c6b.slice/crio-a64e51af19338d566f66a1253bdfc54a2957f685a093ffc1d9626717d822d76b WatchSource:0}: Error finding container a64e51af19338d566f66a1253bdfc54a2957f685a093ffc1d9626717d822d76b: Status 404 returned error can't find the container with id a64e51af19338d566f66a1253bdfc54a2957f685a093ffc1d9626717d822d76b Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.726227 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp"] Dec 05 09:24:10 crc kubenswrapper[4815]: W1205 09:24:10.732143 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9db099f9_ed44_46ed_ab3d_97d09ee0b5b0.slice/crio-09099a319a592858a43f16ab116599cab15f8708a8f9e60bec915aae4a730a33 WatchSource:0}: Error finding container 09099a319a592858a43f16ab116599cab15f8708a8f9e60bec915aae4a730a33: Status 404 returned error can't find the container with id 09099a319a592858a43f16ab116599cab15f8708a8f9e60bec915aae4a730a33 Dec 05 09:24:10 crc kubenswrapper[4815]: E1205 09:24:10.736547 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mck75,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6nl6f_openstack-operators(47664970-bfba-453a-ad06-ac1e03417eac): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 09:24:10 crc kubenswrapper[4815]: E1205 09:24:10.737690 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6cz2x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-g4nbp_openstack-operators(9db099f9-ed44-46ed-ab3d-97d09ee0b5b0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 09:24:10 crc kubenswrapper[4815]: E1205 09:24:10.741126 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mck75,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6nl6f_openstack-operators(47664970-bfba-453a-ad06-ac1e03417eac): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.741197 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f"] Dec 05 09:24:10 crc kubenswrapper[4815]: E1205 09:24:10.742549 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" podUID="47664970-bfba-453a-ad06-ac1e03417eac" Dec 05 09:24:10 crc kubenswrapper[4815]: E1205 09:24:10.750224 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6cz2x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-g4nbp_openstack-operators(9db099f9-ed44-46ed-ab3d-97d09ee0b5b0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 09:24:10 crc kubenswrapper[4815]: E1205 09:24:10.753957 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" podUID="9db099f9-ed44-46ed-ab3d-97d09ee0b5b0" Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.767733 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f"] Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.871034 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws"] Dec 05 09:24:10 crc kubenswrapper[4815]: I1205 09:24:10.886814 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf"] Dec 05 09:24:10 crc kubenswrapper[4815]: E1205 09:24:10.892431 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-94vpt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-gbkbf_openstack-operators(d8b41c8b-030e-4082-98a3-a0e9fc05a208): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 09:24:10 crc kubenswrapper[4815]: E1205 09:24:10.894294 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-94vpt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-gbkbf_openstack-operators(d8b41c8b-030e-4082-98a3-a0e9fc05a208): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 05 09:24:10 crc kubenswrapper[4815]: E1205 09:24:10.896399 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" podUID="d8b41c8b-030e-4082-98a3-a0e9fc05a208" Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.007289 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:11 crc kubenswrapper[4815]: E1205 09:24:11.007561 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.007830 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:11 crc kubenswrapper[4815]: E1205 09:24:11.008001 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs podName:8f0cbf2b-67d6-4058-b173-d24117dfcdd7 nodeName:}" failed. No retries permitted until 2025-12-05 09:24:13.007846351 +0000 UTC m=+1051.886453298 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-2lknm" (UID: "8f0cbf2b-67d6-4058-b173-d24117dfcdd7") : secret "webhook-server-cert" not found Dec 05 09:24:11 crc kubenswrapper[4815]: E1205 09:24:11.008073 4815 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 09:24:11 crc kubenswrapper[4815]: E1205 09:24:11.008162 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs podName:8f0cbf2b-67d6-4058-b173-d24117dfcdd7 nodeName:}" failed. No retries permitted until 2025-12-05 09:24:13.008143799 +0000 UTC m=+1051.886750636 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-2lknm" (UID: "8f0cbf2b-67d6-4058-b173-d24117dfcdd7") : secret "metrics-server-cert" not found Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.122794 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" event={"ID":"9db099f9-ed44-46ed-ab3d-97d09ee0b5b0","Type":"ContainerStarted","Data":"09099a319a592858a43f16ab116599cab15f8708a8f9e60bec915aae4a730a33"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.124273 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" event={"ID":"d8b41c8b-030e-4082-98a3-a0e9fc05a208","Type":"ContainerStarted","Data":"410a38e05c5298f81fcd29a6d0cdacad2688938f30d69e861f3eb73010d781c7"} Dec 05 09:24:11 crc kubenswrapper[4815]: E1205 09:24:11.126786 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" podUID="9db099f9-ed44-46ed-ab3d-97d09ee0b5b0" Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.126966 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" event={"ID":"b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d","Type":"ContainerStarted","Data":"979d0ee4d9c8605926afa1211edf5d53e0ee6eef4f12d566df841a62cefd3e91"} Dec 05 09:24:11 crc kubenswrapper[4815]: E1205 09:24:11.127857 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" podUID="d8b41c8b-030e-4082-98a3-a0e9fc05a208" Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.131990 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" event={"ID":"6b41117c-2922-4337-b00c-2bebd57a2c6b","Type":"ContainerStarted","Data":"a64e51af19338d566f66a1253bdfc54a2957f685a093ffc1d9626717d822d76b"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.133691 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" event={"ID":"926f3f91-cee1-42ae-888f-73ec7f20dd61","Type":"ContainerStarted","Data":"b7c3811efe62e679a3b2f2adddca11680be5eecf3b1ce3817b3a6030728a3270"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.141565 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" event={"ID":"f23f3b0d-7874-4906-81de-b8d9226082d2","Type":"ContainerStarted","Data":"1a6929962135a76cf4d3b30dcdfec1b0c151e0f5a2cadf45436a8df3d339636f"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.148387 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" event={"ID":"378923e7-18f6-4946-8e86-d1ececca15cc","Type":"ContainerStarted","Data":"5dee247a50f3bda520993e8ba7a7beb5cdbb227987deba87e615b4ae98ab3b47"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.149942 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" event={"ID":"f80008bd-769f-4960-be8f-62894fdd7718","Type":"ContainerStarted","Data":"08bcb75757c73f19b77c997480fc1f329e32450cd77b56fd0ed77b82dded8253"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.151148 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" event={"ID":"7482e78f-1387-4e91-b5d4-419f5164fea5","Type":"ContainerStarted","Data":"e171012bb744249ce2894bf7560b0b5565fbbf691532fbbacdc037f6bbe6a53c"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.152273 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" event={"ID":"48f9720a-5d73-4773-85f9-9c571c9322ee","Type":"ContainerStarted","Data":"aac7188fab57b3c9bdc728641fa7893c2d73f565a9b2d2856158fc331b1c0232"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.153334 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f" event={"ID":"19f20867-5315-4be3-836b-57bb0b501b36","Type":"ContainerStarted","Data":"7e9dc8e84397b2b7bd48f813c0cf04926122e29ceac10c25172be8f3ed6460fd"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.154329 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" event={"ID":"cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb","Type":"ContainerStarted","Data":"5ab6cf61a23cb71e4848c8fbd0b9c962a7f47c419ffe42dc6358c86b822e2dc1"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.156096 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" event={"ID":"3665adde-25ad-4261-be50-ef4e21726b7b","Type":"ContainerStarted","Data":"9e7fe6ab5088b598e432997cee09ae1e54137d82e1c30a37fb6afb7e6e9a6cdf"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.157322 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" event={"ID":"ef1e93b5-9a31-4588-a5e4-78dbe6ef0437","Type":"ContainerStarted","Data":"854b3261b3b66d0d0fdd49ea9e0e9e49370beda8e7bc343cc2a30c9fe18a61ba"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.158986 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws" event={"ID":"888d282d-8aa1-4a36-9286-d00af63950ba","Type":"ContainerStarted","Data":"d9728d67726403459be25d207135a2bdf4f5c27a81052539d411bbed69414c76"} Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.167717 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" event={"ID":"47664970-bfba-453a-ad06-ac1e03417eac","Type":"ContainerStarted","Data":"d03506e36209b01e37b8f0fb521a4bf7712fd7dfaa6d61aac2c6578fe016c96f"} Dec 05 09:24:11 crc kubenswrapper[4815]: E1205 09:24:11.174043 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" podUID="47664970-bfba-453a-ad06-ac1e03417eac" Dec 05 09:24:11 crc kubenswrapper[4815]: I1205 09:24:11.818471 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert\") pod \"infra-operator-controller-manager-57548d458d-9pflh\" (UID: \"4abce90f-180f-4302-aecd-89bf5739757c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:11 crc kubenswrapper[4815]: E1205 09:24:11.818632 4815 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 09:24:11 crc kubenswrapper[4815]: E1205 09:24:11.819005 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert podName:4abce90f-180f-4302-aecd-89bf5739757c nodeName:}" failed. No retries permitted until 2025-12-05 09:24:15.818981 +0000 UTC m=+1054.697587867 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert") pod "infra-operator-controller-manager-57548d458d-9pflh" (UID: "4abce90f-180f-4302-aecd-89bf5739757c") : secret "infra-operator-webhook-server-cert" not found Dec 05 09:24:12 crc kubenswrapper[4815]: E1205 09:24:12.240942 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" podUID="d8b41c8b-030e-4082-98a3-a0e9fc05a208" Dec 05 09:24:12 crc kubenswrapper[4815]: E1205 09:24:12.247656 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" podUID="9db099f9-ed44-46ed-ab3d-97d09ee0b5b0" Dec 05 09:24:12 crc kubenswrapper[4815]: E1205 09:24:12.247731 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" podUID="47664970-bfba-453a-ad06-ac1e03417eac" Dec 05 09:24:12 crc kubenswrapper[4815]: I1205 09:24:12.447480 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f5b4xkh\" (UID: \"b7558b70-107c-48ea-ac22-e42b1bcdf47d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:12 crc kubenswrapper[4815]: E1205 09:24:12.447629 4815 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 09:24:12 crc kubenswrapper[4815]: E1205 09:24:12.447671 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert podName:b7558b70-107c-48ea-ac22-e42b1bcdf47d nodeName:}" failed. No retries permitted until 2025-12-05 09:24:16.447658596 +0000 UTC m=+1055.326265423 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" (UID: "b7558b70-107c-48ea-ac22-e42b1bcdf47d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 09:24:13 crc kubenswrapper[4815]: I1205 09:24:13.072945 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:13 crc kubenswrapper[4815]: I1205 09:24:13.073007 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:13 crc kubenswrapper[4815]: E1205 09:24:13.073127 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 09:24:13 crc kubenswrapper[4815]: E1205 09:24:13.073159 4815 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 09:24:13 crc kubenswrapper[4815]: E1205 09:24:13.073218 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs podName:8f0cbf2b-67d6-4058-b173-d24117dfcdd7 nodeName:}" failed. No retries permitted until 2025-12-05 09:24:17.073201226 +0000 UTC m=+1055.951808063 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-2lknm" (UID: "8f0cbf2b-67d6-4058-b173-d24117dfcdd7") : secret "webhook-server-cert" not found Dec 05 09:24:13 crc kubenswrapper[4815]: E1205 09:24:13.073666 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs podName:8f0cbf2b-67d6-4058-b173-d24117dfcdd7 nodeName:}" failed. No retries permitted until 2025-12-05 09:24:17.073229347 +0000 UTC m=+1055.951836184 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-2lknm" (UID: "8f0cbf2b-67d6-4058-b173-d24117dfcdd7") : secret "metrics-server-cert" not found Dec 05 09:24:15 crc kubenswrapper[4815]: I1205 09:24:15.831041 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert\") pod \"infra-operator-controller-manager-57548d458d-9pflh\" (UID: \"4abce90f-180f-4302-aecd-89bf5739757c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:15 crc kubenswrapper[4815]: E1205 09:24:15.831372 4815 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 05 09:24:15 crc kubenswrapper[4815]: E1205 09:24:15.831544 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert podName:4abce90f-180f-4302-aecd-89bf5739757c nodeName:}" failed. No retries permitted until 2025-12-05 09:24:23.831522611 +0000 UTC m=+1062.710129448 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert") pod "infra-operator-controller-manager-57548d458d-9pflh" (UID: "4abce90f-180f-4302-aecd-89bf5739757c") : secret "infra-operator-webhook-server-cert" not found Dec 05 09:24:16 crc kubenswrapper[4815]: E1205 09:24:16.448522 4815 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 09:24:16 crc kubenswrapper[4815]: E1205 09:24:16.448595 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert podName:b7558b70-107c-48ea-ac22-e42b1bcdf47d nodeName:}" failed. No retries permitted until 2025-12-05 09:24:24.448577947 +0000 UTC m=+1063.327184784 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert") pod "openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" (UID: "b7558b70-107c-48ea-ac22-e42b1bcdf47d") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 05 09:24:16 crc kubenswrapper[4815]: I1205 09:24:16.449627 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f5b4xkh\" (UID: \"b7558b70-107c-48ea-ac22-e42b1bcdf47d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:17 crc kubenswrapper[4815]: I1205 09:24:17.159459 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:17 crc kubenswrapper[4815]: I1205 09:24:17.159806 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:17 crc kubenswrapper[4815]: E1205 09:24:17.159621 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 09:24:17 crc kubenswrapper[4815]: E1205 09:24:17.159967 4815 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 05 09:24:17 crc kubenswrapper[4815]: E1205 09:24:17.160068 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs podName:8f0cbf2b-67d6-4058-b173-d24117dfcdd7 nodeName:}" failed. No retries permitted until 2025-12-05 09:24:25.159987186 +0000 UTC m=+1064.038594023 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-2lknm" (UID: "8f0cbf2b-67d6-4058-b173-d24117dfcdd7") : secret "webhook-server-cert" not found Dec 05 09:24:17 crc kubenswrapper[4815]: E1205 09:24:17.160090 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs podName:8f0cbf2b-67d6-4058-b173-d24117dfcdd7 nodeName:}" failed. No retries permitted until 2025-12-05 09:24:25.160081588 +0000 UTC m=+1064.038688425 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs") pod "openstack-operator-controller-manager-54bdf956c4-2lknm" (UID: "8f0cbf2b-67d6-4058-b173-d24117dfcdd7") : secret "metrics-server-cert" not found Dec 05 09:24:23 crc kubenswrapper[4815]: I1205 09:24:23.885395 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert\") pod \"infra-operator-controller-manager-57548d458d-9pflh\" (UID: \"4abce90f-180f-4302-aecd-89bf5739757c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:23 crc kubenswrapper[4815]: I1205 09:24:23.894349 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4abce90f-180f-4302-aecd-89bf5739757c-cert\") pod \"infra-operator-controller-manager-57548d458d-9pflh\" (UID: \"4abce90f-180f-4302-aecd-89bf5739757c\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:24 crc kubenswrapper[4815]: I1205 09:24:24.184862 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:24:24 crc kubenswrapper[4815]: I1205 09:24:24.492412 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f5b4xkh\" (UID: \"b7558b70-107c-48ea-ac22-e42b1bcdf47d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:24 crc kubenswrapper[4815]: I1205 09:24:24.499261 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b7558b70-107c-48ea-ac22-e42b1bcdf47d-cert\") pod \"openstack-baremetal-operator-controller-manager-55c85496f5b4xkh\" (UID: \"b7558b70-107c-48ea-ac22-e42b1bcdf47d\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:24 crc kubenswrapper[4815]: I1205 09:24:24.523581 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:24:25 crc kubenswrapper[4815]: I1205 09:24:25.201668 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:25 crc kubenswrapper[4815]: E1205 09:24:25.201881 4815 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 05 09:24:25 crc kubenswrapper[4815]: I1205 09:24:25.202016 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:25 crc kubenswrapper[4815]: E1205 09:24:25.202507 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs podName:8f0cbf2b-67d6-4058-b173-d24117dfcdd7 nodeName:}" failed. No retries permitted until 2025-12-05 09:24:41.202463015 +0000 UTC m=+1080.081069882 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs") pod "openstack-operator-controller-manager-54bdf956c4-2lknm" (UID: "8f0cbf2b-67d6-4058-b173-d24117dfcdd7") : secret "webhook-server-cert" not found Dec 05 09:24:25 crc kubenswrapper[4815]: I1205 09:24:25.212500 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-metrics-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:30 crc kubenswrapper[4815]: E1205 09:24:30.726409 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f" Dec 05 09:24:30 crc kubenswrapper[4815]: E1205 09:24:30.727306 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v28xz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-8pt5c_openstack-operators(6b41117c-2922-4337-b00c-2bebd57a2c6b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:31 crc kubenswrapper[4815]: E1205 09:24:31.653793 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 05 09:24:31 crc kubenswrapper[4815]: E1205 09:24:31.655230 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k9xrp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-bm8kq_openstack-operators(3665adde-25ad-4261-be50-ef4e21726b7b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:33 crc kubenswrapper[4815]: E1205 09:24:33.083574 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 05 09:24:33 crc kubenswrapper[4815]: E1205 09:24:33.083888 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fbf6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-5jh5q_openstack-operators(ed463fa2-6734-41b5-a34b-cb83436130fd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:41 crc kubenswrapper[4815]: I1205 09:24:41.260918 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:41 crc kubenswrapper[4815]: I1205 09:24:41.284160 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/8f0cbf2b-67d6-4058-b173-d24117dfcdd7-webhook-certs\") pod \"openstack-operator-controller-manager-54bdf956c4-2lknm\" (UID: \"8f0cbf2b-67d6-4058-b173-d24117dfcdd7\") " pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:41 crc kubenswrapper[4815]: I1205 09:24:41.435723 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-r9dpp" Dec 05 09:24:41 crc kubenswrapper[4815]: I1205 09:24:41.443544 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:24:42 crc kubenswrapper[4815]: E1205 09:24:42.701078 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801" Dec 05 09:24:42 crc kubenswrapper[4815]: E1205 09:24:42.701553 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:1d60701214b39cdb0fa70bbe5710f9b131139a9f4b482c2db4058a04daefb801,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bzr9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-j8cxc_openstack-operators(f023493d-1716-40ac-816d-26067f0019f3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:44 crc kubenswrapper[4815]: E1205 09:24:44.940027 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557" Dec 05 09:24:44 crc kubenswrapper[4815]: E1205 09:24:44.940728 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:0b3fb69f35c151895d3dffd514974a9f9fe1c77c3bca69b78b81efb183cf4557,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cnxrc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-ccd95_openstack-operators(7482e78f-1387-4e91-b5d4-419f5164fea5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:48 crc kubenswrapper[4815]: E1205 09:24:48.308264 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530" Dec 05 09:24:48 crc kubenswrapper[4815]: E1205 09:24:48.308774 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:0f523b7e2fa9e86fef986acf07d0c42d5658c475d565f11eaea926ebffcb6530,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-g9kdm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6c548fd776-7mtff_openstack-operators(ef1e93b5-9a31-4588-a5e4-78dbe6ef0437): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:48 crc kubenswrapper[4815]: E1205 09:24:48.750590 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429" Dec 05 09:24:48 crc kubenswrapper[4815]: E1205 09:24:48.750791 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:c4abfc148600dfa85915f3dc911d988ea2335f26cb6b8d749fe79bfe53e5e429,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zx9xt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-5f64f6f8bb-tmg9c_openstack-operators(f23f3b0d-7874-4906-81de-b8d9226082d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:51 crc kubenswrapper[4815]: E1205 09:24:51.699802 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9" Dec 05 09:24:51 crc kubenswrapper[4815]: E1205 09:24:51.700339 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kwz4g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-2k8wt_openstack-operators(f80008bd-769f-4960-be8f-62894fdd7718): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:52 crc kubenswrapper[4815]: E1205 09:24:52.249696 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809" Dec 05 09:24:52 crc kubenswrapper[4815]: E1205 09:24:52.250964 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:abdb733b01e92ac17f565762f30f1d075b44c16421bd06e557f6bb3c319e1809,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vq7pc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-8blh4_openstack-operators(591bafca-5706-418f-b803-d29d1b0865db): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:52 crc kubenswrapper[4815]: E1205 09:24:52.747194 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea" Dec 05 09:24:52 crc kubenswrapper[4815]: E1205 09:24:52.747514 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:f6059a0fbf031d34dcf086d14ce8c0546caeaee23c5780e90b5037c5feee9fea,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-72vrb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-26tt9_openstack-operators(cafb574b-a659-41e6-89d5-c02851274f73): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:53 crc kubenswrapper[4815]: E1205 09:24:53.206733 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 05 09:24:53 crc kubenswrapper[4815]: E1205 09:24:53.206957 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2s7rs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-mvkws_openstack-operators(888d282d-8aa1-4a36-9286-d00af63950ba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:53 crc kubenswrapper[4815]: E1205 09:24:53.208390 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws" podUID="888d282d-8aa1-4a36-9286-d00af63950ba" Dec 05 09:24:53 crc kubenswrapper[4815]: E1205 09:24:53.519541 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws" podUID="888d282d-8aa1-4a36-9286-d00af63950ba" Dec 05 09:24:53 crc kubenswrapper[4815]: E1205 09:24:53.731012 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:24:53 crc kubenswrapper[4815]: E1205 09:24:53.731173 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v28xz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-8pt5c_openstack-operators(6b41117c-2922-4337-b00c-2bebd57a2c6b): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 05 09:24:53 crc kubenswrapper[4815]: E1205 09:24:53.732381 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" podUID="6b41117c-2922-4337-b00c-2bebd57a2c6b" Dec 05 09:24:53 crc kubenswrapper[4815]: E1205 09:24:53.747654 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59" Dec 05 09:24:53 crc kubenswrapper[4815]: E1205 09:24:53.747853 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:635a4aef9d6f0b799e8ec91333dbb312160c001d05b3c63f614c124e0b67cb59,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jgwvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-5jlqr_openstack-operators(378923e7-18f6-4946-8e86-d1ececca15cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:54 crc kubenswrapper[4815]: I1205 09:24:54.173257 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh"] Dec 05 09:24:56 crc kubenswrapper[4815]: E1205 09:24:56.322362 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94" Dec 05 09:24:56 crc kubenswrapper[4815]: E1205 09:24:56.322871 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-94vpt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-gbkbf_openstack-operators(d8b41c8b-030e-4082-98a3-a0e9fc05a208): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:57 crc kubenswrapper[4815]: E1205 09:24:57.122007 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 05 09:24:57 crc kubenswrapper[4815]: E1205 09:24:57.122251 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6cz2x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-g4nbp_openstack-operators(9db099f9-ed44-46ed-ab3d-97d09ee0b5b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:57 crc kubenswrapper[4815]: E1205 09:24:57.953392 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d" Dec 05 09:24:57 crc kubenswrapper[4815]: E1205 09:24:57.953614 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mck75,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6nl6f_openstack-operators(47664970-bfba-453a-ad06-ac1e03417eac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:24:58 crc kubenswrapper[4815]: E1205 09:24:58.869008 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 05 09:24:58 crc kubenswrapper[4815]: E1205 09:24:58.869227 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pgf4s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-25lnj_openstack-operators(48f9720a-5d73-4773-85f9-9c571c9322ee): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:25:03 crc kubenswrapper[4815]: W1205 09:25:03.918415 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7558b70_107c_48ea_ac22_e42b1bcdf47d.slice/crio-670ee2dfdf914b6bfcc572606c87529f29738f8228fc32af7c21a9d1c8290073 WatchSource:0}: Error finding container 670ee2dfdf914b6bfcc572606c87529f29738f8228fc32af7c21a9d1c8290073: Status 404 returned error can't find the container with id 670ee2dfdf914b6bfcc572606c87529f29738f8228fc32af7c21a9d1c8290073 Dec 05 09:25:04 crc kubenswrapper[4815]: E1205 09:25:04.059039 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 05 09:25:04 crc kubenswrapper[4815]: E1205 09:25:04.059537 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zwzzd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-w7vjm_openstack-operators(cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:25:04 crc kubenswrapper[4815]: I1205 09:25:04.345421 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-9pflh"] Dec 05 09:25:04 crc kubenswrapper[4815]: I1205 09:25:04.600619 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" event={"ID":"4abce90f-180f-4302-aecd-89bf5739757c","Type":"ContainerStarted","Data":"15f56acfc0396d5ab72e6db363bd88654d7beca6fdaf67a00640b36e39325e82"} Dec 05 09:25:04 crc kubenswrapper[4815]: I1205 09:25:04.601717 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" event={"ID":"b7558b70-107c-48ea-ac22-e42b1bcdf47d","Type":"ContainerStarted","Data":"670ee2dfdf914b6bfcc572606c87529f29738f8228fc32af7c21a9d1c8290073"} Dec 05 09:25:05 crc kubenswrapper[4815]: I1205 09:25:05.503696 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm"] Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.542681 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.543168 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fbf6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-5jh5q_openstack-operators(ed463fa2-6734-41b5-a34b-cb83436130fd): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.544579 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" podUID="ed463fa2-6734-41b5-a34b-cb83436130fd" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.556579 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.561729 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-72vrb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-7d9dfd778-26tt9_openstack-operators(cafb574b-a659-41e6-89d5-c02851274f73): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.568965 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" podUID="cafb574b-a659-41e6-89d5-c02851274f73" Dec 05 09:25:05 crc kubenswrapper[4815]: I1205 09:25:05.619927 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" event={"ID":"926f3f91-cee1-42ae-888f-73ec7f20dd61","Type":"ContainerStarted","Data":"b786f19ca0562b1ffb1519b111fd09de19befed3b0e90c648fefc52a0064d905"} Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.620544 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.621191 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cnxrc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-ccd95_openstack-operators(7482e78f-1387-4e91-b5d4-419f5164fea5): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.622294 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" podUID="7482e78f-1387-4e91-b5d4-419f5164fea5" Dec 05 09:25:05 crc kubenswrapper[4815]: I1205 09:25:05.624354 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" event={"ID":"8f0cbf2b-67d6-4058-b173-d24117dfcdd7","Type":"ContainerStarted","Data":"9eeffbac47d4bc1a4d3361b33ceca029f551d995f8bef1928bb4aeebebfee985"} Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.655826 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.656013 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k9xrp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-bm8kq_openstack-operators(3665adde-25ad-4261-be50-ef4e21726b7b): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\": context canceled" logger="UnhandledError" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.657180 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:46ba3f23f1d3fb1440deeb279716e4377e79e61736ec2227270349b9618a0fdd\\\": context canceled\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" podUID="3665adde-25ad-4261-be50-ef4e21726b7b" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.875445 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.875727 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kwz4g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-2k8wt_openstack-operators(f80008bd-769f-4960-be8f-62894fdd7718): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.876949 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" podUID="f80008bd-769f-4960-be8f-62894fdd7718" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.950734 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.950930 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bzr9m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-859b6ccc6-j8cxc_openstack-operators(f023493d-1716-40ac-816d-26067f0019f3): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 05 09:25:05 crc kubenswrapper[4815]: E1205 09:25:05.952553 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" podUID="f023493d-1716-40ac-816d-26067f0019f3" Dec 05 09:25:06 crc kubenswrapper[4815]: I1205 09:25:06.658821 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f" event={"ID":"19f20867-5315-4be3-836b-57bb0b501b36","Type":"ContainerStarted","Data":"7527f7e36b95128475470e05a35f57f828f65b0189858bdb1aef465ec26d448f"} Dec 05 09:25:06 crc kubenswrapper[4815]: I1205 09:25:06.663208 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" event={"ID":"8f0cbf2b-67d6-4058-b173-d24117dfcdd7","Type":"ContainerStarted","Data":"84d1527b309d8f248d8e38a1ffecb1cfb2775cb28f72d52d4fc053beb6d0ccb9"} Dec 05 09:25:06 crc kubenswrapper[4815]: I1205 09:25:06.663461 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:25:06 crc kubenswrapper[4815]: I1205 09:25:06.699208 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" event={"ID":"6b41117c-2922-4337-b00c-2bebd57a2c6b","Type":"ContainerStarted","Data":"20cba189a5cc6aa96c21f14a7e236942f01748686f8b12fb663894353114ec5c"} Dec 05 09:25:06 crc kubenswrapper[4815]: I1205 09:25:06.709512 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" event={"ID":"b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d","Type":"ContainerStarted","Data":"3f7821154ac275d26f7b474b064f8eefaa0bebcd42b839184fffa8a0dd18ed89"} Dec 05 09:25:06 crc kubenswrapper[4815]: I1205 09:25:06.769764 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" podStartSLOduration=58.769733249 podStartE2EDuration="58.769733249s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:25:06.723885669 +0000 UTC m=+1105.602492516" watchObservedRunningTime="2025-12-05 09:25:06.769733249 +0000 UTC m=+1105.648340086" Dec 05 09:25:06 crc kubenswrapper[4815]: E1205 09:25:06.922163 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying layer: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:25:06 crc kubenswrapper[4815]: E1205 09:25:06.922356 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vq7pc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-77987cd8cd-8blh4_openstack-operators(591bafca-5706-418f-b803-d29d1b0865db): ErrImagePull: rpc error: code = Canceled desc = copying layer: context canceled" logger="UnhandledError" Dec 05 09:25:06 crc kubenswrapper[4815]: E1205 09:25:06.923547 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying layer: context canceled\"]" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" podUID="591bafca-5706-418f-b803-d29d1b0865db" Dec 05 09:25:07 crc kubenswrapper[4815]: E1205 09:25:07.848825 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" podUID="ed463fa2-6734-41b5-a34b-cb83436130fd" Dec 05 09:25:08 crc kubenswrapper[4815]: I1205 09:25:08.750618 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" event={"ID":"ed463fa2-6734-41b5-a34b-cb83436130fd","Type":"ContainerStarted","Data":"d1f2aba1a1a211eddf3102b84e65c4ccfb517b492f7e48d20f9e78c7de4ddefb"} Dec 05 09:25:08 crc kubenswrapper[4815]: I1205 09:25:08.751676 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" Dec 05 09:25:08 crc kubenswrapper[4815]: I1205 09:25:08.793168 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" event={"ID":"cafb574b-a659-41e6-89d5-c02851274f73","Type":"ContainerStarted","Data":"8c6b52e098e6e472a55a901938884a1e20c61043733761b91e6abcb209ca9941"} Dec 05 09:25:09 crc kubenswrapper[4815]: E1205 09:25:09.396263 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" podUID="ed463fa2-6734-41b5-a34b-cb83436130fd" Dec 05 09:25:09 crc kubenswrapper[4815]: E1205 09:25:09.396424 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" podUID="cafb574b-a659-41e6-89d5-c02851274f73" Dec 05 09:25:09 crc kubenswrapper[4815]: E1205 09:25:09.396513 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" podUID="3665adde-25ad-4261-be50-ef4e21726b7b" Dec 05 09:25:09 crc kubenswrapper[4815]: I1205 09:25:09.808746 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" event={"ID":"3665adde-25ad-4261-be50-ef4e21726b7b","Type":"ContainerStarted","Data":"cf9ad08e62eaa1b965219b273e6ef72689f67ae5ea8487014b7f19f2f91677e3"} Dec 05 09:25:09 crc kubenswrapper[4815]: I1205 09:25:09.810044 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" Dec 05 09:25:09 crc kubenswrapper[4815]: E1205 09:25:09.814010 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" podUID="cafb574b-a659-41e6-89d5-c02851274f73" Dec 05 09:25:09 crc kubenswrapper[4815]: E1205 09:25:09.814159 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" podUID="3665adde-25ad-4261-be50-ef4e21726b7b" Dec 05 09:25:09 crc kubenswrapper[4815]: E1205 09:25:09.814263 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" podUID="ed463fa2-6734-41b5-a34b-cb83436130fd" Dec 05 09:25:10 crc kubenswrapper[4815]: E1205 09:25:10.824250 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" podUID="3665adde-25ad-4261-be50-ef4e21726b7b" Dec 05 09:25:11 crc kubenswrapper[4815]: I1205 09:25:11.453903 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-54bdf956c4-2lknm" Dec 05 09:25:17 crc kubenswrapper[4815]: I1205 09:25:17.976947 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" Dec 05 09:25:17 crc kubenswrapper[4815]: I1205 09:25:17.982998 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" Dec 05 09:25:18 crc kubenswrapper[4815]: E1205 09:25:18.138778 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7" Dec 05 09:25:18 crc kubenswrapper[4815]: E1205 09:25:18.139013 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{600 -3} {} 600m DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{536870912 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7zvg5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod infra-operator-controller-manager-57548d458d-9pflh_openstack-operators(4abce90f-180f-4302-aecd-89bf5739757c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:25:18 crc kubenswrapper[4815]: I1205 09:25:18.475397 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" Dec 05 09:25:18 crc kubenswrapper[4815]: E1205 09:25:18.714056 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:25:18 crc kubenswrapper[4815]: E1205 09:25:18.714263 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z5hgl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-76cc84c6bb-bnb4c_openstack-operators(926f3f91-cee1-42ae-888f-73ec7f20dd61): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:25:18 crc kubenswrapper[4815]: E1205 09:25:18.715431 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" podUID="926f3f91-cee1-42ae-888f-73ec7f20dd61" Dec 05 09:25:18 crc kubenswrapper[4815]: I1205 09:25:18.829569 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" Dec 05 09:25:18 crc kubenswrapper[4815]: I1205 09:25:18.887383 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" Dec 05 09:25:18 crc kubenswrapper[4815]: I1205 09:25:18.889603 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" Dec 05 09:25:18 crc kubenswrapper[4815]: E1205 09:25:18.901019 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:25:18 crc kubenswrapper[4815]: E1205 09:25:18.901264 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r72wl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-998648c74-l5qqd_openstack-operators(b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:25:18 crc kubenswrapper[4815]: E1205 09:25:18.902466 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" podUID="b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d" Dec 05 09:25:18 crc kubenswrapper[4815]: E1205 09:25:18.972144 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:25:18 crc kubenswrapper[4815]: E1205 09:25:18.972550 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jgwvm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-b6456fdb6-5jlqr_openstack-operators(378923e7-18f6-4946-8e86-d1ececca15cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:25:18 crc kubenswrapper[4815]: E1205 09:25:18.973806 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" podUID="378923e7-18f6-4946-8e86-d1ececca15cc" Dec 05 09:25:19 crc kubenswrapper[4815]: E1205 09:25:19.222591 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 05 09:25:19 crc kubenswrapper[4815]: E1205 09:25:19.223079 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mck75,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-6nl6f_openstack-operators(47664970-bfba-453a-ad06-ac1e03417eac): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:25:19 crc kubenswrapper[4815]: E1205 09:25:19.224881 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" podUID="47664970-bfba-453a-ad06-ac1e03417eac" Dec 05 09:25:19 crc kubenswrapper[4815]: I1205 09:25:19.910587 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" Dec 05 09:25:19 crc kubenswrapper[4815]: I1205 09:25:19.915888 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" Dec 05 09:25:20 crc kubenswrapper[4815]: E1205 09:25:20.138635 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" podUID="cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb" Dec 05 09:25:20 crc kubenswrapper[4815]: E1205 09:25:20.138676 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" podUID="d8b41c8b-030e-4082-98a3-a0e9fc05a208" Dec 05 09:25:20 crc kubenswrapper[4815]: E1205 09:25:20.173911 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" podUID="f23f3b0d-7874-4906-81de-b8d9226082d2" Dec 05 09:25:20 crc kubenswrapper[4815]: E1205 09:25:20.224267 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" podUID="48f9720a-5d73-4773-85f9-9c571c9322ee" Dec 05 09:25:20 crc kubenswrapper[4815]: E1205 09:25:20.288907 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" podUID="9db099f9-ed44-46ed-ab3d-97d09ee0b5b0" Dec 05 09:25:20 crc kubenswrapper[4815]: E1205 09:25:20.633082 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" podUID="ef1e93b5-9a31-4588-a5e4-78dbe6ef0437" Dec 05 09:25:20 crc kubenswrapper[4815]: I1205 09:25:20.929773 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" event={"ID":"48f9720a-5d73-4773-85f9-9c571c9322ee","Type":"ContainerStarted","Data":"3b63cf0726ed685f8795b1aa50cc0470637979c460d33bbd15ec9c5ea35739c1"} Dec 05 09:25:20 crc kubenswrapper[4815]: I1205 09:25:20.951977 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" event={"ID":"f80008bd-769f-4960-be8f-62894fdd7718","Type":"ContainerStarted","Data":"12eb45ab3d027a531a53b3450e2cf37d7ac2c8a8b42987689018d6d0e064f38b"} Dec 05 09:25:20 crc kubenswrapper[4815]: I1205 09:25:20.980713 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" event={"ID":"6b41117c-2922-4337-b00c-2bebd57a2c6b","Type":"ContainerStarted","Data":"81f092a2d83b3c7d597cf0a855307ea981d3916e55d6dc530c62936db656749a"} Dec 05 09:25:20 crc kubenswrapper[4815]: I1205 09:25:20.982040 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:20.993795 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.005562 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-8pt5c" podStartSLOduration=4.437242011 podStartE2EDuration="1m13.005542077s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.701052911 +0000 UTC m=+1049.579659748" lastFinishedPulling="2025-12-05 09:25:19.269352977 +0000 UTC m=+1118.147959814" observedRunningTime="2025-12-05 09:25:21.003833729 +0000 UTC m=+1119.882440566" watchObservedRunningTime="2025-12-05 09:25:21.005542077 +0000 UTC m=+1119.884148914" Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.011733 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" event={"ID":"b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d","Type":"ContainerStarted","Data":"6706683699c62fa85d13e4d0992083908025fe0954c71c8379a6fbad576e686f"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.018776 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" event={"ID":"7482e78f-1387-4e91-b5d4-419f5164fea5","Type":"ContainerStarted","Data":"ac6209924070049217ade1d4f33b24ed56203cfa3d75e6ea6bf393667b2be9c6"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.044728 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" event={"ID":"926f3f91-cee1-42ae-888f-73ec7f20dd61","Type":"ContainerStarted","Data":"69b523a64b6098c78bd880dafebb43a0e75463bdc7c4f14a3c46a7bac43cacf6"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.062838 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" event={"ID":"ef1e93b5-9a31-4588-a5e4-78dbe6ef0437","Type":"ContainerStarted","Data":"8e6be97f045d59c2804a82d6377504cfb270b3243dde31cb95194f8b1d87b2d1"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.071889 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-l5qqd" podStartSLOduration=29.854566386 podStartE2EDuration="1m13.07187549s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.516928741 +0000 UTC m=+1049.395535578" lastFinishedPulling="2025-12-05 09:24:53.734237845 +0000 UTC m=+1092.612844682" observedRunningTime="2025-12-05 09:25:21.066509692 +0000 UTC m=+1119.945116529" watchObservedRunningTime="2025-12-05 09:25:21.07187549 +0000 UTC m=+1119.950482327" Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.088158 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws" event={"ID":"888d282d-8aa1-4a36-9286-d00af63950ba","Type":"ContainerStarted","Data":"bdf246e89d59df602e30b53140b020aa56337b001793c5e2625c3a10dedd014b"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.097026 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" event={"ID":"d8b41c8b-030e-4082-98a3-a0e9fc05a208","Type":"ContainerStarted","Data":"60fee4aad4f04d0c819aa1a4731081ca548851c0212c030e7a14a978c4c67ac7"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.110938 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" event={"ID":"9db099f9-ed44-46ed-ab3d-97d09ee0b5b0","Type":"ContainerStarted","Data":"e93ea986879090179ea1f7242cdb79929c684a17f020e8136a278ac5fc61a8e2"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.143186 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" event={"ID":"f023493d-1716-40ac-816d-26067f0019f3","Type":"ContainerStarted","Data":"6deb84569459123d503146b7b7ff57348fb7effb4e8e1c07c149e2db73316540"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.163667 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" event={"ID":"cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb","Type":"ContainerStarted","Data":"f5867c43a35dff2a206056bcc7d7a203a8e92d5762c92b5e56c05529a48d5f69"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.208771 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" event={"ID":"591bafca-5706-418f-b803-d29d1b0865db","Type":"ContainerStarted","Data":"f68b2a667eb56c2520e5f85c6b28c9031d907009667ff823907feb821a24293e"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.223745 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" event={"ID":"b7558b70-107c-48ea-ac22-e42b1bcdf47d","Type":"ContainerStarted","Data":"c07507204e473235a137a708e72845aa9df890f3481e3fb9ac2dd48cd15b4a0c"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.229373 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-76cc84c6bb-bnb4c" podStartSLOduration=27.502783211 podStartE2EDuration="1m13.229347717s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.550179205 +0000 UTC m=+1049.428786042" lastFinishedPulling="2025-12-05 09:24:56.276743711 +0000 UTC m=+1095.155350548" observedRunningTime="2025-12-05 09:25:21.135337563 +0000 UTC m=+1120.013944400" watchObservedRunningTime="2025-12-05 09:25:21.229347717 +0000 UTC m=+1120.107954574" Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.288278 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" event={"ID":"3665adde-25ad-4261-be50-ef4e21726b7b","Type":"ContainerStarted","Data":"83bae45114cfd52ab6c125a15dc82fda5de2b57c4c01e38e7eac41fb33f6b449"} Dec 05 09:25:21 crc kubenswrapper[4815]: E1205 09:25:21.309762 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" podUID="4abce90f-180f-4302-aecd-89bf5739757c" Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.340426 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f" event={"ID":"19f20867-5315-4be3-836b-57bb0b501b36","Type":"ContainerStarted","Data":"6a496d5734526b7838dea16503e39c08e28fe717c9c87a40c4734cde3ee25921"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.340998 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f" Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.344432 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" event={"ID":"f23f3b0d-7874-4906-81de-b8d9226082d2","Type":"ContainerStarted","Data":"1a9a281e69cb63fda73262409c8bfa4507009578e7ca1f04fb31216b3c58f8c5"} Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.347128 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f" Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.420444 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-mvkws" podStartSLOduration=4.554170721 podStartE2EDuration="1m12.420426387s" podCreationTimestamp="2025-12-05 09:24:09 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.876720407 +0000 UTC m=+1049.755327244" lastFinishedPulling="2025-12-05 09:25:18.742976073 +0000 UTC m=+1117.621582910" observedRunningTime="2025-12-05 09:25:21.375839352 +0000 UTC m=+1120.254446189" watchObservedRunningTime="2025-12-05 09:25:21.420426387 +0000 UTC m=+1120.299033234" Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.484649 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-j4r7f" podStartSLOduration=4.94273886 podStartE2EDuration="1m13.484628531s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.727520538 +0000 UTC m=+1049.606127375" lastFinishedPulling="2025-12-05 09:25:19.269410209 +0000 UTC m=+1118.148017046" observedRunningTime="2025-12-05 09:25:21.472223641 +0000 UTC m=+1120.350830488" watchObservedRunningTime="2025-12-05 09:25:21.484628531 +0000 UTC m=+1120.363235368" Dec 05 09:25:21 crc kubenswrapper[4815]: I1205 09:25:21.622864 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-bm8kq" podStartSLOduration=16.239405042 podStartE2EDuration="1m13.622841899s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.484233102 +0000 UTC m=+1049.362839939" lastFinishedPulling="2025-12-05 09:25:07.867669959 +0000 UTC m=+1106.746276796" observedRunningTime="2025-12-05 09:25:21.616931577 +0000 UTC m=+1120.495538414" watchObservedRunningTime="2025-12-05 09:25:21.622841899 +0000 UTC m=+1120.501448746" Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.388342 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" event={"ID":"4abce90f-180f-4302-aecd-89bf5739757c","Type":"ContainerStarted","Data":"81cabc5a7137cba915ef8cf3a7466b22507b550a5ceedf9768a543771cd3f68e"} Dec 05 09:25:22 crc kubenswrapper[4815]: E1205 09:25:22.393778 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" podUID="4abce90f-180f-4302-aecd-89bf5739757c" Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.396039 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" event={"ID":"ed463fa2-6734-41b5-a34b-cb83436130fd","Type":"ContainerStarted","Data":"bb9b123d30798c84fa022d2835cc9c51b593da18e7aeae94a413b59fc8b293b4"} Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.419899 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" event={"ID":"cafb574b-a659-41e6-89d5-c02851274f73","Type":"ContainerStarted","Data":"3dad0f9eb339a4acb85acd97dca397145c3b7e30c92a8e087081ac31d07f9d15"} Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.444764 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" event={"ID":"f023493d-1716-40ac-816d-26067f0019f3","Type":"ContainerStarted","Data":"fde536d811c877677d6986767816189e94e86c49ba0215849559ddda67b33002"} Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.445473 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.463600 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" event={"ID":"378923e7-18f6-4946-8e86-d1ececca15cc","Type":"ContainerStarted","Data":"abe3398e43be79f81ce6768e6088accd2cf72ff81901440fa462fe2ac864e7fc"} Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.506504 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" event={"ID":"f80008bd-769f-4960-be8f-62894fdd7718","Type":"ContainerStarted","Data":"b3c5572c292bc6af882f922fa956865cc4fa36b3dcdb30c06de38e799c9e6f85"} Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.507098 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.526279 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" event={"ID":"591bafca-5706-418f-b803-d29d1b0865db","Type":"ContainerStarted","Data":"d32ff120c9367da5ddacefa53217b67a46d7b00020553a803fbf517f088e25a4"} Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.526746 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.539840 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" event={"ID":"7482e78f-1387-4e91-b5d4-419f5164fea5","Type":"ContainerStarted","Data":"9dd751651fd143ede0827af4088e37c12d868c70cf5f55e9f35dc5746cd87c7e"} Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.540545 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.543049 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" event={"ID":"b7558b70-107c-48ea-ac22-e42b1bcdf47d","Type":"ContainerStarted","Data":"f0e6ac8f0c8824ec967ed1e799786f9b7d634036297a4263900d5f70c465e83e"} Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.543072 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.664089 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-5jh5q" podStartSLOduration=5.075060218 podStartE2EDuration="1m15.664067412s" podCreationTimestamp="2025-12-05 09:24:07 +0000 UTC" firstStartedPulling="2025-12-05 09:24:09.67625365 +0000 UTC m=+1048.554860487" lastFinishedPulling="2025-12-05 09:25:20.265260844 +0000 UTC m=+1119.143867681" observedRunningTime="2025-12-05 09:25:22.662039566 +0000 UTC m=+1121.540646423" watchObservedRunningTime="2025-12-05 09:25:22.664067412 +0000 UTC m=+1121.542674259" Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.665864 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-26tt9" podStartSLOduration=4.635593301 podStartE2EDuration="1m15.665851611s" podCreationTimestamp="2025-12-05 09:24:07 +0000 UTC" firstStartedPulling="2025-12-05 09:24:09.203388206 +0000 UTC m=+1048.081995043" lastFinishedPulling="2025-12-05 09:25:20.233646516 +0000 UTC m=+1119.112253353" observedRunningTime="2025-12-05 09:25:22.526533942 +0000 UTC m=+1121.405140769" watchObservedRunningTime="2025-12-05 09:25:22.665851611 +0000 UTC m=+1121.544458448" Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.724226 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" podStartSLOduration=6.8728596490000005 podStartE2EDuration="1m15.724194034s" podCreationTimestamp="2025-12-05 09:24:07 +0000 UTC" firstStartedPulling="2025-12-05 09:24:09.864701088 +0000 UTC m=+1048.743307915" lastFinishedPulling="2025-12-05 09:25:18.716035463 +0000 UTC m=+1117.594642300" observedRunningTime="2025-12-05 09:25:22.71313839 +0000 UTC m=+1121.591745227" watchObservedRunningTime="2025-12-05 09:25:22.724194034 +0000 UTC m=+1121.602800871" Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.904177 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" podStartSLOduration=7.455634754 podStartE2EDuration="1m15.90415846s" podCreationTimestamp="2025-12-05 09:24:07 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.425423616 +0000 UTC m=+1049.304030453" lastFinishedPulling="2025-12-05 09:25:18.873947322 +0000 UTC m=+1117.752554159" observedRunningTime="2025-12-05 09:25:22.904099798 +0000 UTC m=+1121.782706625" watchObservedRunningTime="2025-12-05 09:25:22.90415846 +0000 UTC m=+1121.782765287" Dec 05 09:25:22 crc kubenswrapper[4815]: I1205 09:25:22.906288 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" podStartSLOduration=6.941456055 podStartE2EDuration="1m14.906274408s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.547548073 +0000 UTC m=+1049.426154920" lastFinishedPulling="2025-12-05 09:25:18.512366446 +0000 UTC m=+1117.390973273" observedRunningTime="2025-12-05 09:25:22.836951103 +0000 UTC m=+1121.715557930" watchObservedRunningTime="2025-12-05 09:25:22.906274408 +0000 UTC m=+1121.784881245" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.064100 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" podStartSLOduration=60.08236864 podStartE2EDuration="1m15.064060733s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:25:03.930760687 +0000 UTC m=+1102.809367524" lastFinishedPulling="2025-12-05 09:25:18.91245278 +0000 UTC m=+1117.791059617" observedRunningTime="2025-12-05 09:25:23.031805846 +0000 UTC m=+1121.910412673" watchObservedRunningTime="2025-12-05 09:25:23.064060733 +0000 UTC m=+1121.942667580" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.142467 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" podStartSLOduration=6.863980115 podStartE2EDuration="1m16.142444347s" podCreationTimestamp="2025-12-05 09:24:07 +0000 UTC" firstStartedPulling="2025-12-05 09:24:09.595555582 +0000 UTC m=+1048.474162419" lastFinishedPulling="2025-12-05 09:25:18.874019814 +0000 UTC m=+1117.752626651" observedRunningTime="2025-12-05 09:25:23.133103591 +0000 UTC m=+1122.011710438" watchObservedRunningTime="2025-12-05 09:25:23.142444347 +0000 UTC m=+1122.021051194" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.552251 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" event={"ID":"d8b41c8b-030e-4082-98a3-a0e9fc05a208","Type":"ContainerStarted","Data":"ff9de2abfa4c05744750ddf903c59158250e17df047986283e12a6b38f032959"} Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.552549 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.554827 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" event={"ID":"9db099f9-ed44-46ed-ab3d-97d09ee0b5b0","Type":"ContainerStarted","Data":"aeec49246e816e1432f8d554c8455a6967b5c995f2eabdbccd8508141dd8e81f"} Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.555087 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.556416 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" event={"ID":"48f9720a-5d73-4773-85f9-9c571c9322ee","Type":"ContainerStarted","Data":"e7e7085cc766e5fadcff0742244afefdc0bb74cb3336128def0ed98174c41446"} Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.556536 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.558403 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" event={"ID":"cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb","Type":"ContainerStarted","Data":"6a3fa0b556928ecd2c2931cac896b11ba37cf0e58bf24fe1b44a29faba4e7830"} Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.558537 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.560740 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" event={"ID":"ef1e93b5-9a31-4588-a5e4-78dbe6ef0437","Type":"ContainerStarted","Data":"61bd15a09ca02fa157245be3ee3231e5dac020353db6469e91bc6bf455d8af81"} Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.560869 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.563024 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" event={"ID":"f23f3b0d-7874-4906-81de-b8d9226082d2","Type":"ContainerStarted","Data":"3edef071d5a0cb3c0c76485c3c75c05a81661709d7a58c7437f547e2c91e75ce"} Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.563523 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.565798 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" event={"ID":"378923e7-18f6-4946-8e86-d1ececca15cc","Type":"ContainerStarted","Data":"cc485d58eee8e28162baec8530c38d43c7372897c9fe01b84168458fa59d78be"} Dec 05 09:25:23 crc kubenswrapper[4815]: E1205 09:25:23.569132 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/infra-operator@sha256:09a6d0613ee2d3c1c809fc36c22678458ac271e0da87c970aec0a5339f5423f7\\\"\"" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" podUID="4abce90f-180f-4302-aecd-89bf5739757c" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.596372 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" podStartSLOduration=4.508005334 podStartE2EDuration="1m15.59634796s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.892281825 +0000 UTC m=+1049.770888662" lastFinishedPulling="2025-12-05 09:25:21.980624451 +0000 UTC m=+1120.859231288" observedRunningTime="2025-12-05 09:25:23.591089575 +0000 UTC m=+1122.469696412" watchObservedRunningTime="2025-12-05 09:25:23.59634796 +0000 UTC m=+1122.474954807" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.626329 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" podStartSLOduration=5.373852697 podStartE2EDuration="1m15.626309233s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.542426782 +0000 UTC m=+1049.421033619" lastFinishedPulling="2025-12-05 09:25:20.794883318 +0000 UTC m=+1119.673490155" observedRunningTime="2025-12-05 09:25:23.623463505 +0000 UTC m=+1122.502070352" watchObservedRunningTime="2025-12-05 09:25:23.626309233 +0000 UTC m=+1122.504916070" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.685972 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" podStartSLOduration=5.2742782009999996 podStartE2EDuration="1m16.685951562s" podCreationTimestamp="2025-12-05 09:24:07 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.389397956 +0000 UTC m=+1049.268004793" lastFinishedPulling="2025-12-05 09:25:21.801071317 +0000 UTC m=+1120.679678154" observedRunningTime="2025-12-05 09:25:23.683228897 +0000 UTC m=+1122.561835734" watchObservedRunningTime="2025-12-05 09:25:23.685951562 +0000 UTC m=+1122.564558399" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.724371 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" podStartSLOduration=4.511784488 podStartE2EDuration="1m15.724349467s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.54272271 +0000 UTC m=+1049.421329547" lastFinishedPulling="2025-12-05 09:25:21.755287689 +0000 UTC m=+1120.633894526" observedRunningTime="2025-12-05 09:25:23.719036251 +0000 UTC m=+1122.597643098" watchObservedRunningTime="2025-12-05 09:25:23.724349467 +0000 UTC m=+1122.602956304" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.767461 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" podStartSLOduration=4.999539141 podStartE2EDuration="1m16.767443701s" podCreationTimestamp="2025-12-05 09:24:07 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.402744313 +0000 UTC m=+1049.281351150" lastFinishedPulling="2025-12-05 09:25:22.170648873 +0000 UTC m=+1121.049255710" observedRunningTime="2025-12-05 09:25:23.758557047 +0000 UTC m=+1122.637163884" watchObservedRunningTime="2025-12-05 09:25:23.767443701 +0000 UTC m=+1122.646050538" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.798371 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" podStartSLOduration=4.977431625 podStartE2EDuration="1m16.798352791s" podCreationTimestamp="2025-12-05 09:24:07 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.157748871 +0000 UTC m=+1049.036355708" lastFinishedPulling="2025-12-05 09:25:21.978670037 +0000 UTC m=+1120.857276874" observedRunningTime="2025-12-05 09:25:23.795342768 +0000 UTC m=+1122.673949605" watchObservedRunningTime="2025-12-05 09:25:23.798352791 +0000 UTC m=+1122.676959628" Dec 05 09:25:23 crc kubenswrapper[4815]: I1205 09:25:23.954930 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" podStartSLOduration=5.89213188 podStartE2EDuration="1m16.954909023s" podCreationTimestamp="2025-12-05 09:24:07 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.737544344 +0000 UTC m=+1049.616151181" lastFinishedPulling="2025-12-05 09:25:21.800321487 +0000 UTC m=+1120.678928324" observedRunningTime="2025-12-05 09:25:23.953940156 +0000 UTC m=+1122.832547003" watchObservedRunningTime="2025-12-05 09:25:23.954909023 +0000 UTC m=+1122.833515860" Dec 05 09:25:24 crc kubenswrapper[4815]: I1205 09:25:24.579323 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" Dec 05 09:25:28 crc kubenswrapper[4815]: I1205 09:25:28.006055 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-j8cxc" Dec 05 09:25:28 crc kubenswrapper[4815]: I1205 09:25:28.153645 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-tmg9c" Dec 05 09:25:28 crc kubenswrapper[4815]: I1205 09:25:28.371572 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-8blh4" Dec 05 09:25:28 crc kubenswrapper[4815]: I1205 09:25:28.394582 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-g4nbp" Dec 05 09:25:28 crc kubenswrapper[4815]: I1205 09:25:28.609311 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-7mtff" Dec 05 09:25:28 crc kubenswrapper[4815]: I1205 09:25:28.618853 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-w7vjm" Dec 05 09:25:28 crc kubenswrapper[4815]: I1205 09:25:28.672138 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-2k8wt" Dec 05 09:25:28 crc kubenswrapper[4815]: I1205 09:25:28.708829 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-ccd95" Dec 05 09:25:28 crc kubenswrapper[4815]: I1205 09:25:28.776873 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-25lnj" Dec 05 09:25:28 crc kubenswrapper[4815]: I1205 09:25:28.828371 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-5jlqr" Dec 05 09:25:29 crc kubenswrapper[4815]: I1205 09:25:29.360858 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-gbkbf" Dec 05 09:25:33 crc kubenswrapper[4815]: I1205 09:25:33.643062 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" event={"ID":"47664970-bfba-453a-ad06-ac1e03417eac","Type":"ContainerStarted","Data":"ed7c6fb2cfd3d112ed163bc85b241aa6b2e622adbb19a81bfaeabc36efae1bdb"} Dec 05 09:25:33 crc kubenswrapper[4815]: I1205 09:25:33.643419 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" event={"ID":"47664970-bfba-453a-ad06-ac1e03417eac","Type":"ContainerStarted","Data":"ab25948ab1fd7f691866273729a37757b62ca2265fd8229f510097cf7484c882"} Dec 05 09:25:33 crc kubenswrapper[4815]: I1205 09:25:33.643658 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" Dec 05 09:25:33 crc kubenswrapper[4815]: I1205 09:25:33.661226 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" podStartSLOduration=3.72184967 podStartE2EDuration="1m25.661212342s" podCreationTimestamp="2025-12-05 09:24:08 +0000 UTC" firstStartedPulling="2025-12-05 09:24:10.736380621 +0000 UTC m=+1049.614987448" lastFinishedPulling="2025-12-05 09:25:32.675743283 +0000 UTC m=+1131.554350120" observedRunningTime="2025-12-05 09:25:33.659068233 +0000 UTC m=+1132.537675070" watchObservedRunningTime="2025-12-05 09:25:33.661212342 +0000 UTC m=+1132.539819169" Dec 05 09:25:34 crc kubenswrapper[4815]: I1205 09:25:34.533278 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" Dec 05 09:25:36 crc kubenswrapper[4815]: I1205 09:25:36.663483 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" event={"ID":"4abce90f-180f-4302-aecd-89bf5739757c","Type":"ContainerStarted","Data":"7eb771c3f5b8a3e984154837aa21e4bed3aa1cbf7686fd7e2ea76609481ed67e"} Dec 05 09:25:36 crc kubenswrapper[4815]: I1205 09:25:36.664856 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:25:36 crc kubenswrapper[4815]: I1205 09:25:36.682585 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" podStartSLOduration=58.27776213 podStartE2EDuration="1m29.682563216s" podCreationTimestamp="2025-12-05 09:24:07 +0000 UTC" firstStartedPulling="2025-12-05 09:25:04.464165974 +0000 UTC m=+1103.342772821" lastFinishedPulling="2025-12-05 09:25:35.86896707 +0000 UTC m=+1134.747573907" observedRunningTime="2025-12-05 09:25:36.679107022 +0000 UTC m=+1135.557713869" watchObservedRunningTime="2025-12-05 09:25:36.682563216 +0000 UTC m=+1135.561170063" Dec 05 09:25:38 crc kubenswrapper[4815]: I1205 09:25:38.955243 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-6nl6f" Dec 05 09:25:44 crc kubenswrapper[4815]: I1205 09:25:44.190435 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" Dec 05 09:25:50 crc kubenswrapper[4815]: I1205 09:25:50.192907 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:25:50 crc kubenswrapper[4815]: I1205 09:25:50.193584 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.183144 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-t25mz"] Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.184755 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.188075 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.188309 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.188515 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-8r58v" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.192544 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.212113 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-t25mz"] Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.303074 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-567c455747-56fgz"] Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.304370 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.308775 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.328307 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567c455747-56fgz"] Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.385257 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d27d5139-5739-4d26-8273-2b78e96ab006-config\") pod \"dnsmasq-dns-5cd484bb89-t25mz\" (UID: \"d27d5139-5739-4d26-8273-2b78e96ab006\") " pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.385341 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp6mj\" (UniqueName: \"kubernetes.io/projected/d27d5139-5739-4d26-8273-2b78e96ab006-kube-api-access-mp6mj\") pod \"dnsmasq-dns-5cd484bb89-t25mz\" (UID: \"d27d5139-5739-4d26-8273-2b78e96ab006\") " pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.487078 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-config\") pod \"dnsmasq-dns-567c455747-56fgz\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.487138 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d27d5139-5739-4d26-8273-2b78e96ab006-config\") pod \"dnsmasq-dns-5cd484bb89-t25mz\" (UID: \"d27d5139-5739-4d26-8273-2b78e96ab006\") " pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.487342 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5j8r\" (UniqueName: \"kubernetes.io/projected/2cadca3c-b906-4fee-afff-548df9cd6584-kube-api-access-r5j8r\") pod \"dnsmasq-dns-567c455747-56fgz\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.487388 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-dns-svc\") pod \"dnsmasq-dns-567c455747-56fgz\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.487413 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp6mj\" (UniqueName: \"kubernetes.io/projected/d27d5139-5739-4d26-8273-2b78e96ab006-kube-api-access-mp6mj\") pod \"dnsmasq-dns-5cd484bb89-t25mz\" (UID: \"d27d5139-5739-4d26-8273-2b78e96ab006\") " pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.488033 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d27d5139-5739-4d26-8273-2b78e96ab006-config\") pod \"dnsmasq-dns-5cd484bb89-t25mz\" (UID: \"d27d5139-5739-4d26-8273-2b78e96ab006\") " pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.518978 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp6mj\" (UniqueName: \"kubernetes.io/projected/d27d5139-5739-4d26-8273-2b78e96ab006-kube-api-access-mp6mj\") pod \"dnsmasq-dns-5cd484bb89-t25mz\" (UID: \"d27d5139-5739-4d26-8273-2b78e96ab006\") " pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.588379 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-config\") pod \"dnsmasq-dns-567c455747-56fgz\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.588459 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5j8r\" (UniqueName: \"kubernetes.io/projected/2cadca3c-b906-4fee-afff-548df9cd6584-kube-api-access-r5j8r\") pod \"dnsmasq-dns-567c455747-56fgz\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.588537 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-dns-svc\") pod \"dnsmasq-dns-567c455747-56fgz\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.589317 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-dns-svc\") pod \"dnsmasq-dns-567c455747-56fgz\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.589929 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-config\") pod \"dnsmasq-dns-567c455747-56fgz\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.606888 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5j8r\" (UniqueName: \"kubernetes.io/projected/2cadca3c-b906-4fee-afff-548df9cd6584-kube-api-access-r5j8r\") pod \"dnsmasq-dns-567c455747-56fgz\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.688504 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:26:00 crc kubenswrapper[4815]: I1205 09:26:00.816105 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" Dec 05 09:26:01 crc kubenswrapper[4815]: I1205 09:26:01.183260 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-567c455747-56fgz"] Dec 05 09:26:01 crc kubenswrapper[4815]: I1205 09:26:01.274845 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-t25mz"] Dec 05 09:26:01 crc kubenswrapper[4815]: I1205 09:26:01.836363 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" event={"ID":"d27d5139-5739-4d26-8273-2b78e96ab006","Type":"ContainerStarted","Data":"2d53f5299e9e7cd6e8aa24a1b5d81844455979dc84cb4bbfab14288e45589c2f"} Dec 05 09:26:01 crc kubenswrapper[4815]: I1205 09:26:01.840426 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-56fgz" event={"ID":"2cadca3c-b906-4fee-afff-548df9cd6584","Type":"ContainerStarted","Data":"73f448533d5e5bf73ff2e98e7059d1e820c55241beef178aeff3458a8a541e64"} Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.395623 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-56fgz"] Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.445046 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-sd2gz"] Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.446283 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.462456 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-sd2gz"] Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.545618 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-dns-svc\") pod \"dnsmasq-dns-bc4b48fc9-sd2gz\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.545689 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlcn5\" (UniqueName: \"kubernetes.io/projected/f8ea7e34-94d1-49a1-942e-90f4b446dab6-kube-api-access-dlcn5\") pod \"dnsmasq-dns-bc4b48fc9-sd2gz\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.545748 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-config\") pod \"dnsmasq-dns-bc4b48fc9-sd2gz\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.647132 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlcn5\" (UniqueName: \"kubernetes.io/projected/f8ea7e34-94d1-49a1-942e-90f4b446dab6-kube-api-access-dlcn5\") pod \"dnsmasq-dns-bc4b48fc9-sd2gz\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.647208 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-config\") pod \"dnsmasq-dns-bc4b48fc9-sd2gz\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.647312 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-dns-svc\") pod \"dnsmasq-dns-bc4b48fc9-sd2gz\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.648372 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-dns-svc\") pod \"dnsmasq-dns-bc4b48fc9-sd2gz\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.649679 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-config\") pod \"dnsmasq-dns-bc4b48fc9-sd2gz\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.706609 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlcn5\" (UniqueName: \"kubernetes.io/projected/f8ea7e34-94d1-49a1-942e-90f4b446dab6-kube-api-access-dlcn5\") pod \"dnsmasq-dns-bc4b48fc9-sd2gz\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.790159 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.795737 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-t25mz"] Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.818407 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb666b895-f72xc"] Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.826447 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.837926 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-f72xc"] Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.952764 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmmc7\" (UniqueName: \"kubernetes.io/projected/d9b960cd-8a61-4cbe-b456-f11fc47345dc-kube-api-access-tmmc7\") pod \"dnsmasq-dns-cb666b895-f72xc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.952909 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-dns-svc\") pod \"dnsmasq-dns-cb666b895-f72xc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:26:03 crc kubenswrapper[4815]: I1205 09:26:03.953065 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-config\") pod \"dnsmasq-dns-cb666b895-f72xc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.056449 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-dns-svc\") pod \"dnsmasq-dns-cb666b895-f72xc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.056616 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-config\") pod \"dnsmasq-dns-cb666b895-f72xc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.056645 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmmc7\" (UniqueName: \"kubernetes.io/projected/d9b960cd-8a61-4cbe-b456-f11fc47345dc-kube-api-access-tmmc7\") pod \"dnsmasq-dns-cb666b895-f72xc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.057767 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-dns-svc\") pod \"dnsmasq-dns-cb666b895-f72xc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.059366 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-config\") pod \"dnsmasq-dns-cb666b895-f72xc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.080959 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmmc7\" (UniqueName: \"kubernetes.io/projected/d9b960cd-8a61-4cbe-b456-f11fc47345dc-kube-api-access-tmmc7\") pod \"dnsmasq-dns-cb666b895-f72xc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.171270 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.525551 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-sd2gz"] Dec 05 09:26:04 crc kubenswrapper[4815]: W1205 09:26:04.533047 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8ea7e34_94d1_49a1_942e_90f4b446dab6.slice/crio-9d71f4739f8dc294fb7a7a8af31274a6ba7fadca12309e5d38438f8421bb943c WatchSource:0}: Error finding container 9d71f4739f8dc294fb7a7a8af31274a6ba7fadca12309e5d38438f8421bb943c: Status 404 returned error can't find the container with id 9d71f4739f8dc294fb7a7a8af31274a6ba7fadca12309e5d38438f8421bb943c Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.598861 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.602542 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.607805 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.608132 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.608447 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.608528 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.608630 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-628cb" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.608756 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.608811 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.638627 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.770900 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-f72xc"] Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.773298 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.773378 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.773648 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.773683 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd2b7\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-kube-api-access-gd2b7\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.773704 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.773759 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.774996 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.775023 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.775112 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-config-data\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.775195 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.775244 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: W1205 09:26:04.811595 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9b960cd_8a61_4cbe_b456_f11fc47345dc.slice/crio-807bed6e5fda3f9b2af3094f688c778e3e634bcff5d8a48b489e1cbb264b332e WatchSource:0}: Error finding container 807bed6e5fda3f9b2af3094f688c778e3e634bcff5d8a48b489e1cbb264b332e: Status 404 returned error can't find the container with id 807bed6e5fda3f9b2af3094f688c778e3e634bcff5d8a48b489e1cbb264b332e Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.878777 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.878822 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd2b7\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-kube-api-access-gd2b7\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.878837 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.878891 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.878910 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.879464 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.879512 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.879542 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.879579 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-config-data\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.879606 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.879647 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.880255 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-config-data\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.880270 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.880313 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.880337 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.880760 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.882512 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-server-conf\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.885191 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.885359 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.888965 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.893105 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-pod-info\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.900448 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd2b7\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-kube-api-access-gd2b7\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.920809 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" event={"ID":"f8ea7e34-94d1-49a1-942e-90f4b446dab6","Type":"ContainerStarted","Data":"9d71f4739f8dc294fb7a7a8af31274a6ba7fadca12309e5d38438f8421bb943c"} Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.925904 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-f72xc" event={"ID":"d9b960cd-8a61-4cbe-b456-f11fc47345dc","Type":"ContainerStarted","Data":"807bed6e5fda3f9b2af3094f688c778e3e634bcff5d8a48b489e1cbb264b332e"} Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.933721 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.938270 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.967800 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.969797 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.980376 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.980397 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.980603 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.980729 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.981316 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.981467 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.981647 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-lmbsv" Dec 05 09:26:04 crc kubenswrapper[4815]: I1205 09:26:04.987595 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.090223 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.090416 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.090475 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.090526 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.090552 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.090586 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.090617 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/852e4372-ed4f-4350-b013-ab36741cd980-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.090644 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.090692 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.090719 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhrl5\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-kube-api-access-xhrl5\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.090806 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/852e4372-ed4f-4350-b013-ab36741cd980-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.191959 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/852e4372-ed4f-4350-b013-ab36741cd980-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.192315 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.192352 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.192385 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.192415 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.192437 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.192465 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.192508 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/852e4372-ed4f-4350-b013-ab36741cd980-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.192535 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.192577 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.192602 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhrl5\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-kube-api-access-xhrl5\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.192825 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.193025 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.193279 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.193376 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.193954 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.198285 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.198996 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.203992 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/852e4372-ed4f-4350-b013-ab36741cd980-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.204023 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/852e4372-ed4f-4350-b013-ab36741cd980-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.210188 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhrl5\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-kube-api-access-xhrl5\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.234561 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.247735 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.313297 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.540392 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.842547 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.935294 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e965503d-d60a-4ce1-9dee-b74fd5eb66a0","Type":"ContainerStarted","Data":"bd3a035bbc6129e455bf025a5966fc590afe5550e0a0aff06e14d588aa280624"} Dec 05 09:26:05 crc kubenswrapper[4815]: I1205 09:26:05.936656 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"852e4372-ed4f-4350-b013-ab36741cd980","Type":"ContainerStarted","Data":"cb6b03b36955480bbb604abd7fa7a3fb73ca932bec9eb06df9a4eb0574fd1506"} Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.354624 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.356265 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.365576 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-l5fhm" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.365688 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.365594 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.365710 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.369475 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.380650 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.517408 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.517528 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.517598 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-config-data-default\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.517619 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.517668 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-kolla-config\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.517692 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.517723 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rz5b2\" (UniqueName: \"kubernetes.io/projected/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-kube-api-access-rz5b2\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.517766 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.619556 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.619716 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.619758 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-config-data-default\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.619780 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.619811 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-kolla-config\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.619825 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.619854 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rz5b2\" (UniqueName: \"kubernetes.io/projected/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-kube-api-access-rz5b2\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.619880 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.620268 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.621150 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.628368 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-kolla-config\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.630089 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.631326 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-config-data-default\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.644326 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.645566 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.651158 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rz5b2\" (UniqueName: \"kubernetes.io/projected/ae4a421a-317a-4a4b-bcb1-2717aba4d25f-kube-api-access-rz5b2\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.693426 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"openstack-galera-0\" (UID: \"ae4a421a-317a-4a4b-bcb1-2717aba4d25f\") " pod="openstack/openstack-galera-0" Dec 05 09:26:06 crc kubenswrapper[4815]: I1205 09:26:06.985159 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.771948 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.773466 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.779079 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.779122 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-m8mnn" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.779456 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.780740 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.793909 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.886972 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.887031 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.887062 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.887133 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zl69\" (UniqueName: \"kubernetes.io/projected/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-kube-api-access-9zl69\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.887162 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.887184 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.887202 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.887225 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.988201 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.988252 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.988330 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zl69\" (UniqueName: \"kubernetes.io/projected/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-kube-api-access-9zl69\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.988359 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.988376 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.988389 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.988411 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.988433 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.989782 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.989815 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.990157 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.991178 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.991671 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.995048 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:07 crc kubenswrapper[4815]: I1205 09:26:07.995802 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.021137 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zl69\" (UniqueName: \"kubernetes.io/projected/8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1-kube-api-access-9zl69\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.039960 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-cell1-galera-0\" (UID: \"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1\") " pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.088099 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.089071 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.101175 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.116777 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.124482 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.127944 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.128217 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-298gl" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.190929 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/28f58061-0d04-4513-96f0-7d221940a154-memcached-tls-certs\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.191045 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/28f58061-0d04-4513-96f0-7d221940a154-kolla-config\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.191118 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28f58061-0d04-4513-96f0-7d221940a154-combined-ca-bundle\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.191143 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28f58061-0d04-4513-96f0-7d221940a154-config-data\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.191167 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jtw9l\" (UniqueName: \"kubernetes.io/projected/28f58061-0d04-4513-96f0-7d221940a154-kube-api-access-jtw9l\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.292528 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/28f58061-0d04-4513-96f0-7d221940a154-kolla-config\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.292626 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28f58061-0d04-4513-96f0-7d221940a154-combined-ca-bundle\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.292649 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28f58061-0d04-4513-96f0-7d221940a154-config-data\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.292671 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jtw9l\" (UniqueName: \"kubernetes.io/projected/28f58061-0d04-4513-96f0-7d221940a154-kube-api-access-jtw9l\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.292714 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/28f58061-0d04-4513-96f0-7d221940a154-memcached-tls-certs\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.300746 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/28f58061-0d04-4513-96f0-7d221940a154-memcached-tls-certs\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.301881 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/28f58061-0d04-4513-96f0-7d221940a154-kolla-config\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.302219 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/28f58061-0d04-4513-96f0-7d221940a154-config-data\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.307311 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28f58061-0d04-4513-96f0-7d221940a154-combined-ca-bundle\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.321032 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jtw9l\" (UniqueName: \"kubernetes.io/projected/28f58061-0d04-4513-96f0-7d221940a154-kube-api-access-jtw9l\") pod \"memcached-0\" (UID: \"28f58061-0d04-4513-96f0-7d221940a154\") " pod="openstack/memcached-0" Dec 05 09:26:08 crc kubenswrapper[4815]: I1205 09:26:08.415453 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 05 09:26:09 crc kubenswrapper[4815]: I1205 09:26:09.008070 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 05 09:26:09 crc kubenswrapper[4815]: W1205 09:26:09.077593 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae4a421a_317a_4a4b_bcb1_2717aba4d25f.slice/crio-279689f28885cf5f365770444c8a407370f1d9c08b04e563a640002a53d39b35 WatchSource:0}: Error finding container 279689f28885cf5f365770444c8a407370f1d9c08b04e563a640002a53d39b35: Status 404 returned error can't find the container with id 279689f28885cf5f365770444c8a407370f1d9c08b04e563a640002a53d39b35 Dec 05 09:26:09 crc kubenswrapper[4815]: I1205 09:26:09.079959 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 05 09:26:09 crc kubenswrapper[4815]: I1205 09:26:09.100744 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 05 09:26:09 crc kubenswrapper[4815]: W1205 09:26:09.106006 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e67fbbb_c9b8_46c1_b761_c76a3f05dfe1.slice/crio-402f1de2e86c14a4448a4fb7525a57e0af02e4d585e9ea8ce2d0dfb4f25816b9 WatchSource:0}: Error finding container 402f1de2e86c14a4448a4fb7525a57e0af02e4d585e9ea8ce2d0dfb4f25816b9: Status 404 returned error can't find the container with id 402f1de2e86c14a4448a4fb7525a57e0af02e4d585e9ea8ce2d0dfb4f25816b9 Dec 05 09:26:09 crc kubenswrapper[4815]: W1205 09:26:09.124065 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28f58061_0d04_4513_96f0_7d221940a154.slice/crio-f9bb70ec3db137c035a3bdb6e9287e1a5ba01105146d35ffe1393ccbc9c6703a WatchSource:0}: Error finding container f9bb70ec3db137c035a3bdb6e9287e1a5ba01105146d35ffe1393ccbc9c6703a: Status 404 returned error can't find the container with id f9bb70ec3db137c035a3bdb6e9287e1a5ba01105146d35ffe1393ccbc9c6703a Dec 05 09:26:09 crc kubenswrapper[4815]: I1205 09:26:09.987865 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1","Type":"ContainerStarted","Data":"402f1de2e86c14a4448a4fb7525a57e0af02e4d585e9ea8ce2d0dfb4f25816b9"} Dec 05 09:26:09 crc kubenswrapper[4815]: I1205 09:26:09.996250 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae4a421a-317a-4a4b-bcb1-2717aba4d25f","Type":"ContainerStarted","Data":"279689f28885cf5f365770444c8a407370f1d9c08b04e563a640002a53d39b35"} Dec 05 09:26:10 crc kubenswrapper[4815]: I1205 09:26:10.000604 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"28f58061-0d04-4513-96f0-7d221940a154","Type":"ContainerStarted","Data":"f9bb70ec3db137c035a3bdb6e9287e1a5ba01105146d35ffe1393ccbc9c6703a"} Dec 05 09:26:10 crc kubenswrapper[4815]: I1205 09:26:10.062102 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:26:10 crc kubenswrapper[4815]: I1205 09:26:10.063210 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 09:26:10 crc kubenswrapper[4815]: I1205 09:26:10.064467 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:26:10 crc kubenswrapper[4815]: I1205 09:26:10.068868 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-jcn7r" Dec 05 09:26:10 crc kubenswrapper[4815]: I1205 09:26:10.124385 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dmc7\" (UniqueName: \"kubernetes.io/projected/96590e06-205c-41dd-9fa3-ee99f499510c-kube-api-access-6dmc7\") pod \"kube-state-metrics-0\" (UID: \"96590e06-205c-41dd-9fa3-ee99f499510c\") " pod="openstack/kube-state-metrics-0" Dec 05 09:26:10 crc kubenswrapper[4815]: I1205 09:26:10.230208 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dmc7\" (UniqueName: \"kubernetes.io/projected/96590e06-205c-41dd-9fa3-ee99f499510c-kube-api-access-6dmc7\") pod \"kube-state-metrics-0\" (UID: \"96590e06-205c-41dd-9fa3-ee99f499510c\") " pod="openstack/kube-state-metrics-0" Dec 05 09:26:10 crc kubenswrapper[4815]: I1205 09:26:10.306589 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dmc7\" (UniqueName: \"kubernetes.io/projected/96590e06-205c-41dd-9fa3-ee99f499510c-kube-api-access-6dmc7\") pod \"kube-state-metrics-0\" (UID: \"96590e06-205c-41dd-9fa3-ee99f499510c\") " pod="openstack/kube-state-metrics-0" Dec 05 09:26:10 crc kubenswrapper[4815]: I1205 09:26:10.402572 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 09:26:10 crc kubenswrapper[4815]: I1205 09:26:10.899669 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.309090 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-v5mcb"] Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.315985 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.329624 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-wsz9s" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.334046 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.358508 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-v5mcb"] Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.396086 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.425811 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-ovn-controller-tls-certs\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.425891 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-var-run-ovn\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.425929 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-var-run\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.425955 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-var-log-ovn\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.426007 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-combined-ca-bundle\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.426068 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-scripts\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.426106 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2czrl\" (UniqueName: \"kubernetes.io/projected/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-kube-api-access-2czrl\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.456569 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-zlv59"] Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.458605 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.465451 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zlv59"] Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.527811 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wgw8h\" (UniqueName: \"kubernetes.io/projected/6e811382-f977-41d3-aa18-65fe948261dd-kube-api-access-wgw8h\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.527875 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-var-run-ovn\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.528011 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-var-run\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.528050 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-var-run\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.528073 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-var-log-ovn\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.528123 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e811382-f977-41d3-aa18-65fe948261dd-scripts\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.528169 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-combined-ca-bundle\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.528242 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-var-log\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.528271 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-etc-ovs\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.528290 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-scripts\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.528316 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-var-lib\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.528337 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2czrl\" (UniqueName: \"kubernetes.io/projected/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-kube-api-access-2czrl\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.528394 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-ovn-controller-tls-certs\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.529422 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-var-run-ovn\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.529809 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-var-run\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.530199 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-var-log-ovn\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.532699 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-scripts\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.536682 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-combined-ca-bundle\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.545358 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-ovn-controller-tls-certs\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.556634 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2czrl\" (UniqueName: \"kubernetes.io/projected/0e0fc690-fb2a-4381-93df-1adcb21c0c7d-kube-api-access-2czrl\") pod \"ovn-controller-v5mcb\" (UID: \"0e0fc690-fb2a-4381-93df-1adcb21c0c7d\") " pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.629911 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e811382-f977-41d3-aa18-65fe948261dd-scripts\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.629990 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-var-log\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.630027 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-etc-ovs\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.630064 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-var-lib\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.630119 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wgw8h\" (UniqueName: \"kubernetes.io/projected/6e811382-f977-41d3-aa18-65fe948261dd-kube-api-access-wgw8h\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.630157 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-var-run\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.630289 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-var-run\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.632244 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e811382-f977-41d3-aa18-65fe948261dd-scripts\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.632336 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-var-log\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.632429 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-etc-ovs\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.632541 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/6e811382-f977-41d3-aa18-65fe948261dd-var-lib\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.648512 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v5mcb" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.650704 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wgw8h\" (UniqueName: \"kubernetes.io/projected/6e811382-f977-41d3-aa18-65fe948261dd-kube-api-access-wgw8h\") pod \"ovn-controller-ovs-zlv59\" (UID: \"6e811382-f977-41d3-aa18-65fe948261dd\") " pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:13 crc kubenswrapper[4815]: I1205 09:26:13.779177 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.217455 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.219433 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.223371 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.223399 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.232099 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.233283 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.234816 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.235437 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-sfbzw" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.353542 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.353636 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.353712 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.353736 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-config\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.353860 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.353939 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.353969 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.354107 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfd5q\" (UniqueName: \"kubernetes.io/projected/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-kube-api-access-tfd5q\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.456160 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.456258 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.456289 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.456353 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfd5q\" (UniqueName: \"kubernetes.io/projected/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-kube-api-access-tfd5q\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.456401 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.456438 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.456481 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.456565 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-config\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.457338 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.458757 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.461938 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-config\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.477508 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.477528 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.485387 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.487347 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfd5q\" (UniqueName: \"kubernetes.io/projected/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-kube-api-access-tfd5q\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.488431 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a065bb70-93f8-4e26-8ce0-6f92b7d7cba6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.514972 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6\") " pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:14 crc kubenswrapper[4815]: I1205 09:26:14.554245 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.617422 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.620606 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.624116 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.624230 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.624128 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-97hml" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.624844 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.628065 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.730535 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82494398-26f8-416f-88cd-65a7365cd2c2-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.730662 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/82494398-26f8-416f-88cd-65a7365cd2c2-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.730683 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82494398-26f8-416f-88cd-65a7365cd2c2-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.730749 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/82494398-26f8-416f-88cd-65a7365cd2c2-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.730803 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82494398-26f8-416f-88cd-65a7365cd2c2-config\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.730829 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pqn2\" (UniqueName: \"kubernetes.io/projected/82494398-26f8-416f-88cd-65a7365cd2c2-kube-api-access-2pqn2\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.730901 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.730984 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/82494398-26f8-416f-88cd-65a7365cd2c2-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.832572 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/82494398-26f8-416f-88cd-65a7365cd2c2-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.832648 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82494398-26f8-416f-88cd-65a7365cd2c2-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.832719 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/82494398-26f8-416f-88cd-65a7365cd2c2-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.832768 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82494398-26f8-416f-88cd-65a7365cd2c2-config\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.832798 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pqn2\" (UniqueName: \"kubernetes.io/projected/82494398-26f8-416f-88cd-65a7365cd2c2-kube-api-access-2pqn2\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.832846 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.832934 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/82494398-26f8-416f-88cd-65a7365cd2c2-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.833040 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82494398-26f8-416f-88cd-65a7365cd2c2-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.833325 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/82494398-26f8-416f-88cd-65a7365cd2c2-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.833995 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.834687 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/82494398-26f8-416f-88cd-65a7365cd2c2-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.834880 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/82494398-26f8-416f-88cd-65a7365cd2c2-config\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.837936 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/82494398-26f8-416f-88cd-65a7365cd2c2-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.840056 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82494398-26f8-416f-88cd-65a7365cd2c2-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.843529 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/82494398-26f8-416f-88cd-65a7365cd2c2-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.854183 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pqn2\" (UniqueName: \"kubernetes.io/projected/82494398-26f8-416f-88cd-65a7365cd2c2-kube-api-access-2pqn2\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.864128 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-sb-0\" (UID: \"82494398-26f8-416f-88cd-65a7365cd2c2\") " pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:17 crc kubenswrapper[4815]: W1205 09:26:17.880207 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod96590e06_205c_41dd_9fa3_ee99f499510c.slice/crio-541cadfa3b0512e3021ee18d532d3baede1c0887cc7a9b131292b18b09cddc1f WatchSource:0}: Error finding container 541cadfa3b0512e3021ee18d532d3baede1c0887cc7a9b131292b18b09cddc1f: Status 404 returned error can't find the container with id 541cadfa3b0512e3021ee18d532d3baede1c0887cc7a9b131292b18b09cddc1f Dec 05 09:26:17 crc kubenswrapper[4815]: I1205 09:26:17.952000 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 05 09:26:18 crc kubenswrapper[4815]: I1205 09:26:18.104605 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"96590e06-205c-41dd-9fa3-ee99f499510c","Type":"ContainerStarted","Data":"541cadfa3b0512e3021ee18d532d3baede1c0887cc7a9b131292b18b09cddc1f"} Dec 05 09:26:20 crc kubenswrapper[4815]: I1205 09:26:20.192603 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:26:20 crc kubenswrapper[4815]: I1205 09:26:20.192673 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:26:34 crc kubenswrapper[4815]: I1205 09:26:25.565713 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" podUID="b7558b70-107c-48ea-ac22-e42b1bcdf47d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.84:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:26:39 crc kubenswrapper[4815]: E1205 09:26:39.956972 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a" Dec 05 09:26:39 crc kubenswrapper[4815]: E1205 09:26:39.957745 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rz5b2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(ae4a421a-317a-4a4b-bcb1-2717aba4d25f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:26:39 crc kubenswrapper[4815]: E1205 09:26:39.958965 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="ae4a421a-317a-4a4b-bcb1-2717aba4d25f" Dec 05 09:26:40 crc kubenswrapper[4815]: E1205 09:26:40.335115 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a\\\"\"" pod="openstack/openstack-galera-0" podUID="ae4a421a-317a-4a4b-bcb1-2717aba4d25f" Dec 05 09:26:41 crc kubenswrapper[4815]: E1205 09:26:41.136472 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a" Dec 05 09:26:41 crc kubenswrapper[4815]: E1205 09:26:41.136887 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9zl69,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-cell1-galera-0_openstack(8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:26:41 crc kubenswrapper[4815]: E1205 09:26:41.140628 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-cell1-galera-0" podUID="8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1" Dec 05 09:26:41 crc kubenswrapper[4815]: E1205 09:26:41.342686 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-mariadb@sha256:d9a2d8b19d3de4658dd26a2e781d00002e937738bfe6a1d0cf6c68c015085f4a\\\"\"" pod="openstack/openstack-cell1-galera-0" podUID="8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1" Dec 05 09:26:41 crc kubenswrapper[4815]: E1205 09:26:41.398063 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d" Dec 05 09:26:41 crc kubenswrapper[4815]: E1205 09:26:41.398258 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xhrl5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(852e4372-ed4f-4350-b013-ab36741cd980): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:26:41 crc kubenswrapper[4815]: E1205 09:26:41.400269 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="852e4372-ed4f-4350-b013-ab36741cd980" Dec 05 09:26:42 crc kubenswrapper[4815]: E1205 09:26:42.359166 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq@sha256:c2e607504d3c90c201f29ca74f0ee51de1ef2c5444d597d2ff3160ab9b3ce84d\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="852e4372-ed4f-4350-b013-ab36741cd980" Dec 05 09:26:50 crc kubenswrapper[4815]: I1205 09:26:50.192042 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:26:50 crc kubenswrapper[4815]: I1205 09:26:50.193051 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:26:50 crc kubenswrapper[4815]: I1205 09:26:50.193108 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:26:50 crc kubenswrapper[4815]: I1205 09:26:50.194119 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e422e582a96b96c6d83e3c0fc7b930a3d237267c29a704a2f4827be174a6d09c"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:26:50 crc kubenswrapper[4815]: I1205 09:26:50.194186 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://e422e582a96b96c6d83e3c0fc7b930a3d237267c29a704a2f4827be174a6d09c" gracePeriod=600 Dec 05 09:26:52 crc kubenswrapper[4815]: I1205 09:26:52.922318 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zlv59"] Dec 05 09:26:53 crc kubenswrapper[4815]: I1205 09:26:53.556171 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="e422e582a96b96c6d83e3c0fc7b930a3d237267c29a704a2f4827be174a6d09c" exitCode=0 Dec 05 09:26:53 crc kubenswrapper[4815]: I1205 09:26:53.556376 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"e422e582a96b96c6d83e3c0fc7b930a3d237267c29a704a2f4827be174a6d09c"} Dec 05 09:26:53 crc kubenswrapper[4815]: I1205 09:26:53.556815 4815 scope.go:117] "RemoveContainer" containerID="bda5cae8332591142b5d1ee8c164d1bfba69b735e07e595565abef36f7e9229e" Dec 05 09:26:58 crc kubenswrapper[4815]: E1205 09:26:58.304374 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached@sha256:dad2336390cae6705133deefaa09c9e39512cf29133aa009006e3962c8022108" Dec 05 09:26:58 crc kubenswrapper[4815]: E1205 09:26:58.304823 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached@sha256:dad2336390cae6705133deefaa09c9e39512cf29133aa009006e3962c8022108,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n675h644h5hfdh5d9h6fh5cfh56hf5hc6h8ch5bbh7fh654h5cbh6ch546h5b8hdfh87hb4h589h5h59chch79h66hbch674h5dh5cfh579q,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jtw9l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(28f58061-0d04-4513-96f0-7d221940a154): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:26:58 crc kubenswrapper[4815]: E1205 09:26:58.306723 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="28f58061-0d04-4513-96f0-7d221940a154" Dec 05 09:26:58 crc kubenswrapper[4815]: E1205 09:26:58.604102 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached@sha256:dad2336390cae6705133deefaa09c9e39512cf29133aa009006e3962c8022108\\\"\"" pod="openstack/memcached-0" podUID="28f58061-0d04-4513-96f0-7d221940a154" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.029834 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.029959 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r5j8r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-567c455747-56fgz_openstack(2cadca3c-b906-4fee-afff-548df9cd6584): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.031556 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-567c455747-56fgz" podUID="2cadca3c-b906-4fee-afff-548df9cd6584" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.031730 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.031836 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dlcn5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-bc4b48fc9-sd2gz_openstack(f8ea7e34-94d1-49a1-942e-90f4b446dab6): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.033037 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" podUID="f8ea7e34-94d1-49a1-942e-90f4b446dab6" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.043457 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.043642 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tmmc7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-cb666b895-f72xc_openstack(d9b960cd-8a61-4cbe-b456-f11fc47345dc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.047795 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-cb666b895-f72xc" podUID="d9b960cd-8a61-4cbe-b456-f11fc47345dc" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.052008 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.052181 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mp6mj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5cd484bb89-t25mz_openstack(d27d5139-5739-4d26-8273-2b78e96ab006): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.053254 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" podUID="d27d5139-5739-4d26-8273-2b78e96ab006" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.540866 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-v5mcb"] Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.611938 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.626399 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zlv59" event={"ID":"6e811382-f977-41d3-aa18-65fe948261dd","Type":"ContainerStarted","Data":"e296c6ba96b4ebc22bd37f3ea8bac56a82160f30d81a6b439741d759f30bfbd4"} Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.629180 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792\\\"\"" pod="openstack/dnsmasq-dns-cb666b895-f72xc" podUID="d9b960cd-8a61-4cbe-b456-f11fc47345dc" Dec 05 09:26:59 crc kubenswrapper[4815]: E1205 09:26:59.630022 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:42f5663a161307156673f86e5eaad59f842a4bf25824f48008e69ab18e4ba792\\\"\"" pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" podUID="f8ea7e34-94d1-49a1-942e-90f4b446dab6" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.666618 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-6q6fd"] Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.694523 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.703585 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-6q6fd"] Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.708473 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.834421 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-ovs-rundir\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.834516 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-config\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.834554 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-combined-ca-bundle\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.834573 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2r5l\" (UniqueName: \"kubernetes.io/projected/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-kube-api-access-c2r5l\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.834614 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-ovn-rundir\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.834643 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.896078 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.936455 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-ovn-rundir\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.936521 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.936581 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-ovs-rundir\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.936653 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-config\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.936683 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-combined-ca-bundle\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.936700 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2r5l\" (UniqueName: \"kubernetes.io/projected/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-kube-api-access-c2r5l\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.937682 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-ovs-rundir\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.937756 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-ovn-rundir\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.939861 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-config\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.947529 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.966135 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-combined-ca-bundle\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:26:59 crc kubenswrapper[4815]: I1205 09:26:59.972938 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-sd2gz"] Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:26:59.998213 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2r5l\" (UniqueName: \"kubernetes.io/projected/0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0-kube-api-access-c2r5l\") pod \"ovn-controller-metrics-6q6fd\" (UID: \"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0\") " pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.000766 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-6xt9l"] Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.005662 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.024551 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.031007 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-6xt9l"] Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.036940 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-6q6fd" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.139462 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9wffk\" (UniqueName: \"kubernetes.io/projected/92c86c07-aae3-45a0-b2fd-aa4532f2b542-kube-api-access-9wffk\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.139667 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-ovsdbserver-nb\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.139721 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-dns-svc\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.139774 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-config\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.244187 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9wffk\" (UniqueName: \"kubernetes.io/projected/92c86c07-aae3-45a0-b2fd-aa4532f2b542-kube-api-access-9wffk\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.244263 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-ovsdbserver-nb\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.244318 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-dns-svc\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.244374 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-config\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.245219 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-ovsdbserver-nb\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.294037 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-dns-svc\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.294708 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-config\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.297966 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9wffk\" (UniqueName: \"kubernetes.io/projected/92c86c07-aae3-45a0-b2fd-aa4532f2b542-kube-api-access-9wffk\") pod \"dnsmasq-dns-57db9b5bc9-6xt9l\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: W1205 09:27:00.349514 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e0fc690_fb2a_4381_93df_1adcb21c0c7d.slice/crio-2bd7e30027ff9142064dde7a42cd8569f6561aa7d57c9c22aa99047396e21f5a WatchSource:0}: Error finding container 2bd7e30027ff9142064dde7a42cd8569f6561aa7d57c9c22aa99047396e21f5a: Status 404 returned error can't find the container with id 2bd7e30027ff9142064dde7a42cd8569f6561aa7d57c9c22aa99047396e21f5a Dec 05 09:27:00 crc kubenswrapper[4815]: W1205 09:27:00.359314 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82494398_26f8_416f_88cd_65a7365cd2c2.slice/crio-3a613c378be10c0b15453a4959399d4c9aa429c2698c69f0508be97356b855f9 WatchSource:0}: Error finding container 3a613c378be10c0b15453a4959399d4c9aa429c2698c69f0508be97356b855f9: Status 404 returned error can't find the container with id 3a613c378be10c0b15453a4959399d4c9aa429c2698c69f0508be97356b855f9 Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.374450 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.468683 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.470990 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-f72xc"] Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.478464 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.543807 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-mvfhk"] Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.545142 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.556100 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d27d5139-5739-4d26-8273-2b78e96ab006-config\") pod \"d27d5139-5739-4d26-8273-2b78e96ab006\" (UID: \"d27d5139-5739-4d26-8273-2b78e96ab006\") " Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.556186 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mp6mj\" (UniqueName: \"kubernetes.io/projected/d27d5139-5739-4d26-8273-2b78e96ab006-kube-api-access-mp6mj\") pod \"d27d5139-5739-4d26-8273-2b78e96ab006\" (UID: \"d27d5139-5739-4d26-8273-2b78e96ab006\") " Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.556244 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r5j8r\" (UniqueName: \"kubernetes.io/projected/2cadca3c-b906-4fee-afff-548df9cd6584-kube-api-access-r5j8r\") pod \"2cadca3c-b906-4fee-afff-548df9cd6584\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.556293 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-dns-svc\") pod \"2cadca3c-b906-4fee-afff-548df9cd6584\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.556351 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-config\") pod \"2cadca3c-b906-4fee-afff-548df9cd6584\" (UID: \"2cadca3c-b906-4fee-afff-548df9cd6584\") " Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.558510 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2cadca3c-b906-4fee-afff-548df9cd6584" (UID: "2cadca3c-b906-4fee-afff-548df9cd6584"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.558868 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d27d5139-5739-4d26-8273-2b78e96ab006-config" (OuterVolumeSpecName: "config") pod "d27d5139-5739-4d26-8273-2b78e96ab006" (UID: "d27d5139-5739-4d26-8273-2b78e96ab006"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.559089 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-config" (OuterVolumeSpecName: "config") pod "2cadca3c-b906-4fee-afff-548df9cd6584" (UID: "2cadca3c-b906-4fee-afff-548df9cd6584"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:00 crc kubenswrapper[4815]: W1205 09:27:00.566258 4815 reflector.go:561] object-"openstack"/"ovsdbserver-sb": failed to list *v1.ConfigMap: configmaps "ovsdbserver-sb" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Dec 05 09:27:00 crc kubenswrapper[4815]: E1205 09:27:00.566295 4815 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"ovsdbserver-sb\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovsdbserver-sb\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.578724 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cadca3c-b906-4fee-afff-548df9cd6584-kube-api-access-r5j8r" (OuterVolumeSpecName: "kube-api-access-r5j8r") pod "2cadca3c-b906-4fee-afff-548df9cd6584" (UID: "2cadca3c-b906-4fee-afff-548df9cd6584"). InnerVolumeSpecName "kube-api-access-r5j8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.581895 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d27d5139-5739-4d26-8273-2b78e96ab006-kube-api-access-mp6mj" (OuterVolumeSpecName: "kube-api-access-mp6mj") pod "d27d5139-5739-4d26-8273-2b78e96ab006" (UID: "d27d5139-5739-4d26-8273-2b78e96ab006"). InnerVolumeSpecName "kube-api-access-mp6mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.616195 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-mvfhk"] Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.658284 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46zjb\" (UniqueName: \"kubernetes.io/projected/f38535a2-7fda-4ebc-9d88-241703ec1a1b-kube-api-access-46zjb\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.658339 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-dns-svc\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.658405 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-config\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.658427 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.658483 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.658545 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.658555 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cadca3c-b906-4fee-afff-548df9cd6584-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.658564 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d27d5139-5739-4d26-8273-2b78e96ab006-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.658572 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mp6mj\" (UniqueName: \"kubernetes.io/projected/d27d5139-5739-4d26-8273-2b78e96ab006-kube-api-access-mp6mj\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.658581 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r5j8r\" (UniqueName: \"kubernetes.io/projected/2cadca3c-b906-4fee-afff-548df9cd6584-kube-api-access-r5j8r\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.687283 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v5mcb" event={"ID":"0e0fc690-fb2a-4381-93df-1adcb21c0c7d","Type":"ContainerStarted","Data":"2bd7e30027ff9142064dde7a42cd8569f6561aa7d57c9c22aa99047396e21f5a"} Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.702793 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6","Type":"ContainerStarted","Data":"572eed1f8cff61cad394db4b2ee87ad0cbf526aa96f3d3af86ffaeb7a026c878"} Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.716850 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"82494398-26f8-416f-88cd-65a7365cd2c2","Type":"ContainerStarted","Data":"3a613c378be10c0b15453a4959399d4c9aa429c2698c69f0508be97356b855f9"} Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.737718 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-567c455747-56fgz" event={"ID":"2cadca3c-b906-4fee-afff-548df9cd6584","Type":"ContainerDied","Data":"73f448533d5e5bf73ff2e98e7059d1e820c55241beef178aeff3458a8a541e64"} Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.737824 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-567c455747-56fgz" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.751760 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" event={"ID":"d27d5139-5739-4d26-8273-2b78e96ab006","Type":"ContainerDied","Data":"2d53f5299e9e7cd6e8aa24a1b5d81844455979dc84cb4bbfab14288e45589c2f"} Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.751902 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5cd484bb89-t25mz" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.760735 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.760818 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.760849 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46zjb\" (UniqueName: \"kubernetes.io/projected/f38535a2-7fda-4ebc-9d88-241703ec1a1b-kube-api-access-46zjb\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.760879 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-dns-svc\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.760937 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-config\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.761694 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-config\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.761914 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-nb\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.762429 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-dns-svc\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.804628 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46zjb\" (UniqueName: \"kubernetes.io/projected/f38535a2-7fda-4ebc-9d88-241703ec1a1b-kube-api-access-46zjb\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.929132 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-567c455747-56fgz"] Dec 05 09:27:00 crc kubenswrapper[4815]: I1205 09:27:00.950986 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-567c455747-56fgz"] Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.029665 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-t25mz"] Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.048913 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5cd484bb89-t25mz"] Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.552664 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cadca3c-b906-4fee-afff-548df9cd6584" path="/var/lib/kubelet/pods/2cadca3c-b906-4fee-afff-548df9cd6584/volumes" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.556141 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d27d5139-5739-4d26-8273-2b78e96ab006" path="/var/lib/kubelet/pods/d27d5139-5739-4d26-8273-2b78e96ab006/volumes" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.556670 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-6xt9l"] Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.581540 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.659467 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-6q6fd"] Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.712074 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmmc7\" (UniqueName: \"kubernetes.io/projected/d9b960cd-8a61-4cbe-b456-f11fc47345dc-kube-api-access-tmmc7\") pod \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.712196 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-config\") pod \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.712238 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-dns-svc\") pod \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\" (UID: \"d9b960cd-8a61-4cbe-b456-f11fc47345dc\") " Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.712801 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d9b960cd-8a61-4cbe-b456-f11fc47345dc" (UID: "d9b960cd-8a61-4cbe-b456-f11fc47345dc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.718741 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9b960cd-8a61-4cbe-b456-f11fc47345dc-kube-api-access-tmmc7" (OuterVolumeSpecName: "kube-api-access-tmmc7") pod "d9b960cd-8a61-4cbe-b456-f11fc47345dc" (UID: "d9b960cd-8a61-4cbe-b456-f11fc47345dc"). InnerVolumeSpecName "kube-api-access-tmmc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.719452 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-config" (OuterVolumeSpecName: "config") pod "d9b960cd-8a61-4cbe-b456-f11fc47345dc" (UID: "d9b960cd-8a61-4cbe-b456-f11fc47345dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:01 crc kubenswrapper[4815]: E1205 09:27:01.761819 4815 configmap.go:193] Couldn't get configMap openstack/ovsdbserver-sb: failed to sync configmap cache: timed out waiting for the condition Dec 05 09:27:01 crc kubenswrapper[4815]: E1205 09:27:01.761936 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-sb podName:f38535a2-7fda-4ebc-9d88-241703ec1a1b nodeName:}" failed. No retries permitted until 2025-12-05 09:27:02.261891598 +0000 UTC m=+1221.140498435 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "ovsdbserver-sb" (UniqueName: "kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-sb") pod "dnsmasq-dns-db7757ddc-mvfhk" (UID: "f38535a2-7fda-4ebc-9d88-241703ec1a1b") : failed to sync configmap cache: timed out waiting for the condition Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.780190 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb666b895-f72xc" event={"ID":"d9b960cd-8a61-4cbe-b456-f11fc47345dc","Type":"ContainerDied","Data":"807bed6e5fda3f9b2af3094f688c778e3e634bcff5d8a48b489e1cbb264b332e"} Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.780243 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb666b895-f72xc" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.785211 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" event={"ID":"92c86c07-aae3-45a0-b2fd-aa4532f2b542","Type":"ContainerStarted","Data":"14134714fa288b8e781a682eb405ac7cdd8f482e42f1a0452a40fede7d559a35"} Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.786077 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.799863 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"ca66886398a56fe8786c896f7db80c714344ffc63db0667dbec913c482105f33"} Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.802385 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" event={"ID":"f8ea7e34-94d1-49a1-942e-90f4b446dab6","Type":"ContainerDied","Data":"9d71f4739f8dc294fb7a7a8af31274a6ba7fadca12309e5d38438f8421bb943c"} Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.802552 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bc4b48fc9-sd2gz" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.813890 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmmc7\" (UniqueName: \"kubernetes.io/projected/d9b960cd-8a61-4cbe-b456-f11fc47345dc-kube-api-access-tmmc7\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.813925 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.813938 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d9b960cd-8a61-4cbe-b456-f11fc47345dc-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.912651 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-f72xc"] Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.914935 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlcn5\" (UniqueName: \"kubernetes.io/projected/f8ea7e34-94d1-49a1-942e-90f4b446dab6-kube-api-access-dlcn5\") pod \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.915539 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-dns-svc\") pod \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.915627 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-config\") pod \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\" (UID: \"f8ea7e34-94d1-49a1-942e-90f4b446dab6\") " Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.917047 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-config" (OuterVolumeSpecName: "config") pod "f8ea7e34-94d1-49a1-942e-90f4b446dab6" (UID: "f8ea7e34-94d1-49a1-942e-90f4b446dab6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.917458 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f8ea7e34-94d1-49a1-942e-90f4b446dab6" (UID: "f8ea7e34-94d1-49a1-942e-90f4b446dab6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.923344 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cb666b895-f72xc"] Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.924098 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8ea7e34-94d1-49a1-942e-90f4b446dab6-kube-api-access-dlcn5" (OuterVolumeSpecName: "kube-api-access-dlcn5") pod "f8ea7e34-94d1-49a1-942e-90f4b446dab6" (UID: "f8ea7e34-94d1-49a1-942e-90f4b446dab6"). InnerVolumeSpecName "kube-api-access-dlcn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:01 crc kubenswrapper[4815]: I1205 09:27:01.939829 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 05 09:27:02 crc kubenswrapper[4815]: I1205 09:27:02.019762 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlcn5\" (UniqueName: \"kubernetes.io/projected/f8ea7e34-94d1-49a1-942e-90f4b446dab6-kube-api-access-dlcn5\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:02 crc kubenswrapper[4815]: I1205 09:27:02.019801 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:02 crc kubenswrapper[4815]: I1205 09:27:02.019816 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8ea7e34-94d1-49a1-942e-90f4b446dab6-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:02 crc kubenswrapper[4815]: I1205 09:27:02.181787 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-sd2gz"] Dec 05 09:27:02 crc kubenswrapper[4815]: I1205 09:27:02.193407 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bc4b48fc9-sd2gz"] Dec 05 09:27:02 crc kubenswrapper[4815]: I1205 09:27:02.325007 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:02 crc kubenswrapper[4815]: I1205 09:27:02.326086 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-sb\") pod \"dnsmasq-dns-db7757ddc-mvfhk\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:02 crc kubenswrapper[4815]: I1205 09:27:02.399519 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:02 crc kubenswrapper[4815]: I1205 09:27:02.826930 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6q6fd" event={"ID":"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0","Type":"ContainerStarted","Data":"cbd96656ead677965ed1079d8951ca318f6dfeab9ae21fa06bd782f6d40a6bc2"} Dec 05 09:27:03 crc kubenswrapper[4815]: I1205 09:27:03.450095 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9b960cd-8a61-4cbe-b456-f11fc47345dc" path="/var/lib/kubelet/pods/d9b960cd-8a61-4cbe-b456-f11fc47345dc/volumes" Dec 05 09:27:03 crc kubenswrapper[4815]: I1205 09:27:03.451723 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8ea7e34-94d1-49a1-942e-90f4b446dab6" path="/var/lib/kubelet/pods/f8ea7e34-94d1-49a1-942e-90f4b446dab6/volumes" Dec 05 09:27:03 crc kubenswrapper[4815]: I1205 09:27:03.457032 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-mvfhk"] Dec 05 09:27:03 crc kubenswrapper[4815]: I1205 09:27:03.838999 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1","Type":"ContainerStarted","Data":"00905ab27282b49f460a59ab0116e3a3f38d09686b3ae560b6f76f50de9dc790"} Dec 05 09:27:03 crc kubenswrapper[4815]: I1205 09:27:03.841892 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e965503d-d60a-4ce1-9dee-b74fd5eb66a0","Type":"ContainerStarted","Data":"396633accd6c07faf0270b697296080712b2c885fe98668cd948cdf27a9bfbfa"} Dec 05 09:27:03 crc kubenswrapper[4815]: I1205 09:27:03.848612 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"852e4372-ed4f-4350-b013-ab36741cd980","Type":"ContainerStarted","Data":"fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186"} Dec 05 09:27:03 crc kubenswrapper[4815]: I1205 09:27:03.854056 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae4a421a-317a-4a4b-bcb1-2717aba4d25f","Type":"ContainerStarted","Data":"4bba24715dfb0cdd280dfc4932d54a1a7d4f65700262a2f0861fb8a36fa3b23d"} Dec 05 09:27:03 crc kubenswrapper[4815]: I1205 09:27:03.855640 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"96590e06-205c-41dd-9fa3-ee99f499510c","Type":"ContainerStarted","Data":"a47aa43ff263c21233ba4fef44a4e7c8154161b9c3cec737a33b19d04a682aa7"} Dec 05 09:27:03 crc kubenswrapper[4815]: I1205 09:27:03.856163 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 05 09:27:04 crc kubenswrapper[4815]: I1205 09:27:04.046709 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=8.98444696 podStartE2EDuration="54.046691064s" podCreationTimestamp="2025-12-05 09:26:10 +0000 UTC" firstStartedPulling="2025-12-05 09:26:17.884813136 +0000 UTC m=+1176.763419973" lastFinishedPulling="2025-12-05 09:27:02.94705725 +0000 UTC m=+1221.825664077" observedRunningTime="2025-12-05 09:27:04.04101603 +0000 UTC m=+1222.919622867" watchObservedRunningTime="2025-12-05 09:27:04.046691064 +0000 UTC m=+1222.925297901" Dec 05 09:27:05 crc kubenswrapper[4815]: I1205 09:27:05.875771 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" event={"ID":"f38535a2-7fda-4ebc-9d88-241703ec1a1b","Type":"ContainerStarted","Data":"dcdb4a28c86cc7090d555b60bf024153c553f0988c1cd40ad3235f0e2ad80f51"} Dec 05 09:27:05 crc kubenswrapper[4815]: I1205 09:27:05.877405 4815 generic.go:334] "Generic (PLEG): container finished" podID="92c86c07-aae3-45a0-b2fd-aa4532f2b542" containerID="35a1af36a37de14e23c6772c3f8a10fe9ed1b8fdc6366f3afb7f69f07f8cd449" exitCode=0 Dec 05 09:27:05 crc kubenswrapper[4815]: I1205 09:27:05.877443 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" event={"ID":"92c86c07-aae3-45a0-b2fd-aa4532f2b542","Type":"ContainerDied","Data":"35a1af36a37de14e23c6772c3f8a10fe9ed1b8fdc6366f3afb7f69f07f8cd449"} Dec 05 09:27:06 crc kubenswrapper[4815]: I1205 09:27:06.888084 4815 generic.go:334] "Generic (PLEG): container finished" podID="ae4a421a-317a-4a4b-bcb1-2717aba4d25f" containerID="4bba24715dfb0cdd280dfc4932d54a1a7d4f65700262a2f0861fb8a36fa3b23d" exitCode=0 Dec 05 09:27:06 crc kubenswrapper[4815]: I1205 09:27:06.888121 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae4a421a-317a-4a4b-bcb1-2717aba4d25f","Type":"ContainerDied","Data":"4bba24715dfb0cdd280dfc4932d54a1a7d4f65700262a2f0861fb8a36fa3b23d"} Dec 05 09:27:07 crc kubenswrapper[4815]: I1205 09:27:07.898870 4815 generic.go:334] "Generic (PLEG): container finished" podID="8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1" containerID="00905ab27282b49f460a59ab0116e3a3f38d09686b3ae560b6f76f50de9dc790" exitCode=0 Dec 05 09:27:07 crc kubenswrapper[4815]: I1205 09:27:07.899235 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1","Type":"ContainerDied","Data":"00905ab27282b49f460a59ab0116e3a3f38d09686b3ae560b6f76f50de9dc790"} Dec 05 09:27:10 crc kubenswrapper[4815]: I1205 09:27:10.409074 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 05 09:27:10 crc kubenswrapper[4815]: I1205 09:27:10.931286 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"82494398-26f8-416f-88cd-65a7365cd2c2","Type":"ContainerStarted","Data":"275b15c1040185042af8a375c4871141608690386c53679e2a2685af328c0c8e"} Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.939574 4815 generic.go:334] "Generic (PLEG): container finished" podID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" containerID="8fd74d008a79136a2ac470f0d1d17844e69664d5d1c34c645fd8b9fd0c1e51b0" exitCode=0 Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.939625 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" event={"ID":"f38535a2-7fda-4ebc-9d88-241703ec1a1b","Type":"ContainerDied","Data":"8fd74d008a79136a2ac470f0d1d17844e69664d5d1c34c645fd8b9fd0c1e51b0"} Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.943046 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ae4a421a-317a-4a4b-bcb1-2717aba4d25f","Type":"ContainerStarted","Data":"791b1fd5fa928adaf279a8a4baff69cee57011ddafab8a684eb1b636a582a4b1"} Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.945267 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" event={"ID":"92c86c07-aae3-45a0-b2fd-aa4532f2b542","Type":"ContainerStarted","Data":"29e4959df45b3b0e9294e36efb913156e6274678794229bd25128e72e594fdad"} Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.945535 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.948328 4815 generic.go:334] "Generic (PLEG): container finished" podID="6e811382-f977-41d3-aa18-65fe948261dd" containerID="4037c78bca813505aa14744f0968c66765c255bac999663e4663277075ea6663" exitCode=0 Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.948403 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zlv59" event={"ID":"6e811382-f977-41d3-aa18-65fe948261dd","Type":"ContainerDied","Data":"4037c78bca813505aa14744f0968c66765c255bac999663e4663277075ea6663"} Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.950805 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"82494398-26f8-416f-88cd-65a7365cd2c2","Type":"ContainerStarted","Data":"133f763d2e23dfaa8da5e83340a5150be400bf8700d4cef6b3126561987f5404"} Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.952259 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.953706 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-6q6fd" event={"ID":"0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0","Type":"ContainerStarted","Data":"650c7051060a52db0eec64d9eb57eb2b06e5da2d0379f4390616ee43be38c823"} Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.965822 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1","Type":"ContainerStarted","Data":"b2e651e7d076677fe8a27d50cae5b20e27c1b60b787119b677b59cbfa70e3eee"} Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.971878 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v5mcb" event={"ID":"0e0fc690-fb2a-4381-93df-1adcb21c0c7d","Type":"ContainerStarted","Data":"9cf9b375a09455048ad2e1e884aea59e60f5604e277fc238ddc7f8351e46213e"} Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.972396 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-v5mcb" Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.973907 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6","Type":"ContainerStarted","Data":"58ac813af50f625a86262bbb3eba86d61ac513cf725fb96bf028a0516dab407b"} Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.974199 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a065bb70-93f8-4e26-8ce0-6f92b7d7cba6","Type":"ContainerStarted","Data":"ebc5bb4d863c9a3f1302880970e52d6f53a2da5159bc77e439f16b628a7e26ae"} Dec 05 09:27:11 crc kubenswrapper[4815]: I1205 09:27:11.983159 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" podStartSLOduration=11.703219404 podStartE2EDuration="12.983132472s" podCreationTimestamp="2025-12-05 09:26:59 +0000 UTC" firstStartedPulling="2025-12-05 09:27:01.733711161 +0000 UTC m=+1220.612317998" lastFinishedPulling="2025-12-05 09:27:03.013624229 +0000 UTC m=+1221.892231066" observedRunningTime="2025-12-05 09:27:11.982378521 +0000 UTC m=+1230.860985368" watchObservedRunningTime="2025-12-05 09:27:11.983132472 +0000 UTC m=+1230.861739309" Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.022379 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-6q6fd" podStartSLOduration=4.886518504 podStartE2EDuration="13.022359539s" podCreationTimestamp="2025-12-05 09:26:59 +0000 UTC" firstStartedPulling="2025-12-05 09:27:01.893528722 +0000 UTC m=+1220.772135559" lastFinishedPulling="2025-12-05 09:27:10.029369757 +0000 UTC m=+1228.907976594" observedRunningTime="2025-12-05 09:27:12.001291869 +0000 UTC m=+1230.879898746" watchObservedRunningTime="2025-12-05 09:27:12.022359539 +0000 UTC m=+1230.900966376" Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.115907 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=15.834682977 podStartE2EDuration="1m7.1158875s" podCreationTimestamp="2025-12-05 09:26:05 +0000 UTC" firstStartedPulling="2025-12-05 09:26:09.080914943 +0000 UTC m=+1167.959521780" lastFinishedPulling="2025-12-05 09:27:00.362119466 +0000 UTC m=+1219.240726303" observedRunningTime="2025-12-05 09:27:12.078618199 +0000 UTC m=+1230.957225036" watchObservedRunningTime="2025-12-05 09:27:12.1158875 +0000 UTC m=+1230.994494347" Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.132135 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=48.984067347999996 podStartE2EDuration="56.132116319s" podCreationTimestamp="2025-12-05 09:26:16 +0000 UTC" firstStartedPulling="2025-12-05 09:27:00.380756736 +0000 UTC m=+1219.259363573" lastFinishedPulling="2025-12-05 09:27:07.528805707 +0000 UTC m=+1226.407412544" observedRunningTime="2025-12-05 09:27:12.105093167 +0000 UTC m=+1230.983699994" watchObservedRunningTime="2025-12-05 09:27:12.132116319 +0000 UTC m=+1231.010723156" Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.137615 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=51.961687749 podStartE2EDuration="59.137568737s" podCreationTimestamp="2025-12-05 09:26:13 +0000 UTC" firstStartedPulling="2025-12-05 09:27:00.354168905 +0000 UTC m=+1219.232775742" lastFinishedPulling="2025-12-05 09:27:07.530049893 +0000 UTC m=+1226.408656730" observedRunningTime="2025-12-05 09:27:12.137537726 +0000 UTC m=+1231.016144563" watchObservedRunningTime="2025-12-05 09:27:12.137568737 +0000 UTC m=+1231.016175574" Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.175027 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=14.897966301 podStartE2EDuration="1m6.175006773s" podCreationTimestamp="2025-12-05 09:26:06 +0000 UTC" firstStartedPulling="2025-12-05 09:26:09.10876685 +0000 UTC m=+1167.987373687" lastFinishedPulling="2025-12-05 09:27:00.385807322 +0000 UTC m=+1219.264414159" observedRunningTime="2025-12-05 09:27:12.172657244 +0000 UTC m=+1231.051264081" watchObservedRunningTime="2025-12-05 09:27:12.175006773 +0000 UTC m=+1231.053613610" Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.204703 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-v5mcb" podStartSLOduration=52.027385063 podStartE2EDuration="59.204684862s" podCreationTimestamp="2025-12-05 09:26:13 +0000 UTC" firstStartedPulling="2025-12-05 09:27:00.353934958 +0000 UTC m=+1219.232541795" lastFinishedPulling="2025-12-05 09:27:07.531234757 +0000 UTC m=+1226.409841594" observedRunningTime="2025-12-05 09:27:12.203863069 +0000 UTC m=+1231.082469906" watchObservedRunningTime="2025-12-05 09:27:12.204684862 +0000 UTC m=+1231.083291699" Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.952162 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.986414 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zlv59" event={"ID":"6e811382-f977-41d3-aa18-65fe948261dd","Type":"ContainerStarted","Data":"cabdd615a603ceb34d462f3a47267d9dc5fdfbc30a83b46ae567f72caeb82ab5"} Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.986457 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zlv59" event={"ID":"6e811382-f977-41d3-aa18-65fe948261dd","Type":"ContainerStarted","Data":"dd03ef9d560c595a89db4f7c98a350060a03dcfe4a974eb590d28533c0978603"} Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.986876 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.987035 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:27:12 crc kubenswrapper[4815]: I1205 09:27:12.989590 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" event={"ID":"f38535a2-7fda-4ebc-9d88-241703ec1a1b","Type":"ContainerStarted","Data":"6e63a5434a480287d59e55a265a98a2afe7af7dc9da69f0494958e9ecc809fce"} Dec 05 09:27:13 crc kubenswrapper[4815]: I1205 09:27:13.014398 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-zlv59" podStartSLOduration=51.51685546 podStartE2EDuration="1m0.014381075s" podCreationTimestamp="2025-12-05 09:26:13 +0000 UTC" firstStartedPulling="2025-12-05 09:26:59.031281112 +0000 UTC m=+1217.909887949" lastFinishedPulling="2025-12-05 09:27:07.528806717 +0000 UTC m=+1226.407413564" observedRunningTime="2025-12-05 09:27:13.01003908 +0000 UTC m=+1231.888645917" watchObservedRunningTime="2025-12-05 09:27:13.014381075 +0000 UTC m=+1231.892987912" Dec 05 09:27:13 crc kubenswrapper[4815]: I1205 09:27:13.046180 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" podStartSLOduration=13.046157486 podStartE2EDuration="13.046157486s" podCreationTimestamp="2025-12-05 09:27:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:27:13.033021735 +0000 UTC m=+1231.911628582" watchObservedRunningTime="2025-12-05 09:27:13.046157486 +0000 UTC m=+1231.924764333" Dec 05 09:27:13 crc kubenswrapper[4815]: I1205 09:27:13.997257 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"28f58061-0d04-4513-96f0-7d221940a154","Type":"ContainerStarted","Data":"5decd64daf3067c3574eeaf1216ee5724e68569ca20ae4cbf9d9b38c51842101"} Dec 05 09:27:13 crc kubenswrapper[4815]: I1205 09:27:13.998078 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 05 09:27:13 crc kubenswrapper[4815]: I1205 09:27:13.998116 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:14 crc kubenswrapper[4815]: I1205 09:27:14.018943 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.215966594 podStartE2EDuration="1m6.018924324s" podCreationTimestamp="2025-12-05 09:26:08 +0000 UTC" firstStartedPulling="2025-12-05 09:26:09.126880085 +0000 UTC m=+1168.005486922" lastFinishedPulling="2025-12-05 09:27:12.929837815 +0000 UTC m=+1231.808444652" observedRunningTime="2025-12-05 09:27:14.013656001 +0000 UTC m=+1232.892262838" watchObservedRunningTime="2025-12-05 09:27:14.018924324 +0000 UTC m=+1232.897531161" Dec 05 09:27:14 crc kubenswrapper[4815]: I1205 09:27:14.554797 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 05 09:27:14 crc kubenswrapper[4815]: I1205 09:27:14.555050 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 05 09:27:14 crc kubenswrapper[4815]: I1205 09:27:14.597796 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 05 09:27:15 crc kubenswrapper[4815]: I1205 09:27:15.033225 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 05 09:27:15 crc kubenswrapper[4815]: I1205 09:27:15.191265 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.060913 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.243246 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.246171 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.251795 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.251972 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-t5kbn" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.252238 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.253201 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.272997 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.402924 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3e63e92-1982-4055-9df7-db07096a570e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.402984 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9v98\" (UniqueName: \"kubernetes.io/projected/c3e63e92-1982-4055-9df7-db07096a570e-kube-api-access-m9v98\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.403033 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c3e63e92-1982-4055-9df7-db07096a570e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.403064 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3e63e92-1982-4055-9df7-db07096a570e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.403159 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3e63e92-1982-4055-9df7-db07096a570e-scripts\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.403185 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3e63e92-1982-4055-9df7-db07096a570e-config\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.403223 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3e63e92-1982-4055-9df7-db07096a570e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.505002 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3e63e92-1982-4055-9df7-db07096a570e-scripts\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.506087 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c3e63e92-1982-4055-9df7-db07096a570e-scripts\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.506225 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3e63e92-1982-4055-9df7-db07096a570e-config\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.506937 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c3e63e92-1982-4055-9df7-db07096a570e-config\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.507018 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3e63e92-1982-4055-9df7-db07096a570e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.507934 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3e63e92-1982-4055-9df7-db07096a570e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.507980 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9v98\" (UniqueName: \"kubernetes.io/projected/c3e63e92-1982-4055-9df7-db07096a570e-kube-api-access-m9v98\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.508035 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c3e63e92-1982-4055-9df7-db07096a570e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.508074 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3e63e92-1982-4055-9df7-db07096a570e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.508878 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c3e63e92-1982-4055-9df7-db07096a570e-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.514420 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3e63e92-1982-4055-9df7-db07096a570e-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.516161 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3e63e92-1982-4055-9df7-db07096a570e-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.517203 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3e63e92-1982-4055-9df7-db07096a570e-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.541854 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9v98\" (UniqueName: \"kubernetes.io/projected/c3e63e92-1982-4055-9df7-db07096a570e-kube-api-access-m9v98\") pod \"ovn-northd-0\" (UID: \"c3e63e92-1982-4055-9df7-db07096a570e\") " pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.642145 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.986846 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 05 09:27:16 crc kubenswrapper[4815]: I1205 09:27:16.987412 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 05 09:27:17 crc kubenswrapper[4815]: I1205 09:27:17.191319 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 05 09:27:17 crc kubenswrapper[4815]: I1205 09:27:17.402801 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:27:17 crc kubenswrapper[4815]: I1205 09:27:17.430385 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 05 09:27:17 crc kubenswrapper[4815]: I1205 09:27:17.509331 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-6xt9l"] Dec 05 09:27:17 crc kubenswrapper[4815]: I1205 09:27:17.509606 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" podUID="92c86c07-aae3-45a0-b2fd-aa4532f2b542" containerName="dnsmasq-dns" containerID="cri-o://29e4959df45b3b0e9294e36efb913156e6274678794229bd25128e72e594fdad" gracePeriod=10 Dec 05 09:27:17 crc kubenswrapper[4815]: I1205 09:27:17.515375 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:17 crc kubenswrapper[4815]: I1205 09:27:17.732385 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.068312 4815 generic.go:334] "Generic (PLEG): container finished" podID="92c86c07-aae3-45a0-b2fd-aa4532f2b542" containerID="29e4959df45b3b0e9294e36efb913156e6274678794229bd25128e72e594fdad" exitCode=0 Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.068414 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" event={"ID":"92c86c07-aae3-45a0-b2fd-aa4532f2b542","Type":"ContainerDied","Data":"29e4959df45b3b0e9294e36efb913156e6274678794229bd25128e72e594fdad"} Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.068823 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" event={"ID":"92c86c07-aae3-45a0-b2fd-aa4532f2b542","Type":"ContainerDied","Data":"14134714fa288b8e781a682eb405ac7cdd8f482e42f1a0452a40fede7d559a35"} Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.068855 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14134714fa288b8e781a682eb405ac7cdd8f482e42f1a0452a40fede7d559a35" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.071664 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c3e63e92-1982-4055-9df7-db07096a570e","Type":"ContainerStarted","Data":"34975dfd37882d66e97cd9aca25e7330c00533476a1b3884dfee40815e22a6cf"} Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.074427 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.102343 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.102399 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.151073 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-ovsdbserver-nb\") pod \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.151123 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-config\") pod \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.151175 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-dns-svc\") pod \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.151206 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9wffk\" (UniqueName: \"kubernetes.io/projected/92c86c07-aae3-45a0-b2fd-aa4532f2b542-kube-api-access-9wffk\") pod \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.157901 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92c86c07-aae3-45a0-b2fd-aa4532f2b542-kube-api-access-9wffk" (OuterVolumeSpecName: "kube-api-access-9wffk") pod "92c86c07-aae3-45a0-b2fd-aa4532f2b542" (UID: "92c86c07-aae3-45a0-b2fd-aa4532f2b542"). InnerVolumeSpecName "kube-api-access-9wffk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.201952 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "92c86c07-aae3-45a0-b2fd-aa4532f2b542" (UID: "92c86c07-aae3-45a0-b2fd-aa4532f2b542"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:18 crc kubenswrapper[4815]: E1205 09:27:18.249822 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-config podName:92c86c07-aae3-45a0-b2fd-aa4532f2b542 nodeName:}" failed. No retries permitted until 2025-12-05 09:27:18.749785313 +0000 UTC m=+1237.628392150 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-config") pod "92c86c07-aae3-45a0-b2fd-aa4532f2b542" (UID: "92c86c07-aae3-45a0-b2fd-aa4532f2b542") : error deleting /var/lib/kubelet/pods/92c86c07-aae3-45a0-b2fd-aa4532f2b542/volume-subpaths: remove /var/lib/kubelet/pods/92c86c07-aae3-45a0-b2fd-aa4532f2b542/volume-subpaths: no such file or directory Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.250107 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "92c86c07-aae3-45a0-b2fd-aa4532f2b542" (UID: "92c86c07-aae3-45a0-b2fd-aa4532f2b542"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.253443 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.253466 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.253477 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9wffk\" (UniqueName: \"kubernetes.io/projected/92c86c07-aae3-45a0-b2fd-aa4532f2b542-kube-api-access-9wffk\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.418735 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.713197 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-9wd7r"] Dec 05 09:27:18 crc kubenswrapper[4815]: E1205 09:27:18.714011 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c86c07-aae3-45a0-b2fd-aa4532f2b542" containerName="dnsmasq-dns" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.714081 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c86c07-aae3-45a0-b2fd-aa4532f2b542" containerName="dnsmasq-dns" Dec 05 09:27:18 crc kubenswrapper[4815]: E1205 09:27:18.714104 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c86c07-aae3-45a0-b2fd-aa4532f2b542" containerName="init" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.714114 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c86c07-aae3-45a0-b2fd-aa4532f2b542" containerName="init" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.714322 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="92c86c07-aae3-45a0-b2fd-aa4532f2b542" containerName="dnsmasq-dns" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.715349 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9wd7r" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.723265 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9wd7r"] Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.742294 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-b445-account-create-update-84rcr"] Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.745186 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b445-account-create-update-84rcr" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.760649 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-config\") pod \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\" (UID: \"92c86c07-aae3-45a0-b2fd-aa4532f2b542\") " Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.761848 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-config" (OuterVolumeSpecName: "config") pod "92c86c07-aae3-45a0-b2fd-aa4532f2b542" (UID: "92c86c07-aae3-45a0-b2fd-aa4532f2b542"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.783124 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.793176 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b445-account-create-update-84rcr"] Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.844599 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-bb8e-account-create-update-fbh8q"] Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.845669 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bb8e-account-create-update-fbh8q" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.849298 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.853578 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-kf8dq"] Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.854596 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kf8dq" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.859291 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bb8e-account-create-update-fbh8q"] Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.868359 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkzj2\" (UniqueName: \"kubernetes.io/projected/bdc590e0-1c3a-418e-828b-eb43990319de-kube-api-access-kkzj2\") pod \"glance-db-create-9wd7r\" (UID: \"bdc590e0-1c3a-418e-828b-eb43990319de\") " pod="openstack/glance-db-create-9wd7r" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.868426 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-operator-scripts\") pod \"glance-b445-account-create-update-84rcr\" (UID: \"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd\") " pod="openstack/glance-b445-account-create-update-84rcr" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.868568 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdc590e0-1c3a-418e-828b-eb43990319de-operator-scripts\") pod \"glance-db-create-9wd7r\" (UID: \"bdc590e0-1c3a-418e-828b-eb43990319de\") " pod="openstack/glance-db-create-9wd7r" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.868631 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs9f9\" (UniqueName: \"kubernetes.io/projected/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-kube-api-access-xs9f9\") pod \"glance-b445-account-create-update-84rcr\" (UID: \"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd\") " pod="openstack/glance-b445-account-create-update-84rcr" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.868746 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92c86c07-aae3-45a0-b2fd-aa4532f2b542-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.927361 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-kf8dq"] Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.971644 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmrfd\" (UniqueName: \"kubernetes.io/projected/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-kube-api-access-pmrfd\") pod \"placement-db-create-kf8dq\" (UID: \"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e\") " pod="openstack/placement-db-create-kf8dq" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.971723 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0e72c3-db89-4255-9c0b-50734039276b-operator-scripts\") pod \"placement-bb8e-account-create-update-fbh8q\" (UID: \"cf0e72c3-db89-4255-9c0b-50734039276b\") " pod="openstack/placement-bb8e-account-create-update-fbh8q" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.971766 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkzj2\" (UniqueName: \"kubernetes.io/projected/bdc590e0-1c3a-418e-828b-eb43990319de-kube-api-access-kkzj2\") pod \"glance-db-create-9wd7r\" (UID: \"bdc590e0-1c3a-418e-828b-eb43990319de\") " pod="openstack/glance-db-create-9wd7r" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.971793 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-operator-scripts\") pod \"glance-b445-account-create-update-84rcr\" (UID: \"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd\") " pod="openstack/glance-b445-account-create-update-84rcr" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.971824 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w4jt\" (UniqueName: \"kubernetes.io/projected/cf0e72c3-db89-4255-9c0b-50734039276b-kube-api-access-4w4jt\") pod \"placement-bb8e-account-create-update-fbh8q\" (UID: \"cf0e72c3-db89-4255-9c0b-50734039276b\") " pod="openstack/placement-bb8e-account-create-update-fbh8q" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.971872 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-operator-scripts\") pod \"placement-db-create-kf8dq\" (UID: \"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e\") " pod="openstack/placement-db-create-kf8dq" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.971890 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdc590e0-1c3a-418e-828b-eb43990319de-operator-scripts\") pod \"glance-db-create-9wd7r\" (UID: \"bdc590e0-1c3a-418e-828b-eb43990319de\") " pod="openstack/glance-db-create-9wd7r" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.971935 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs9f9\" (UniqueName: \"kubernetes.io/projected/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-kube-api-access-xs9f9\") pod \"glance-b445-account-create-update-84rcr\" (UID: \"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd\") " pod="openstack/glance-b445-account-create-update-84rcr" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.973002 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdc590e0-1c3a-418e-828b-eb43990319de-operator-scripts\") pod \"glance-db-create-9wd7r\" (UID: \"bdc590e0-1c3a-418e-828b-eb43990319de\") " pod="openstack/glance-db-create-9wd7r" Dec 05 09:27:18 crc kubenswrapper[4815]: I1205 09:27:18.973036 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-operator-scripts\") pod \"glance-b445-account-create-update-84rcr\" (UID: \"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd\") " pod="openstack/glance-b445-account-create-update-84rcr" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.002559 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkzj2\" (UniqueName: \"kubernetes.io/projected/bdc590e0-1c3a-418e-828b-eb43990319de-kube-api-access-kkzj2\") pod \"glance-db-create-9wd7r\" (UID: \"bdc590e0-1c3a-418e-828b-eb43990319de\") " pod="openstack/glance-db-create-9wd7r" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.004363 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs9f9\" (UniqueName: \"kubernetes.io/projected/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-kube-api-access-xs9f9\") pod \"glance-b445-account-create-update-84rcr\" (UID: \"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd\") " pod="openstack/glance-b445-account-create-update-84rcr" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.054552 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9wd7r" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.072838 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-operator-scripts\") pod \"placement-db-create-kf8dq\" (UID: \"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e\") " pod="openstack/placement-db-create-kf8dq" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.073100 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmrfd\" (UniqueName: \"kubernetes.io/projected/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-kube-api-access-pmrfd\") pod \"placement-db-create-kf8dq\" (UID: \"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e\") " pod="openstack/placement-db-create-kf8dq" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.073143 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0e72c3-db89-4255-9c0b-50734039276b-operator-scripts\") pod \"placement-bb8e-account-create-update-fbh8q\" (UID: \"cf0e72c3-db89-4255-9c0b-50734039276b\") " pod="openstack/placement-bb8e-account-create-update-fbh8q" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.073225 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w4jt\" (UniqueName: \"kubernetes.io/projected/cf0e72c3-db89-4255-9c0b-50734039276b-kube-api-access-4w4jt\") pod \"placement-bb8e-account-create-update-fbh8q\" (UID: \"cf0e72c3-db89-4255-9c0b-50734039276b\") " pod="openstack/placement-bb8e-account-create-update-fbh8q" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.074133 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-operator-scripts\") pod \"placement-db-create-kf8dq\" (UID: \"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e\") " pod="openstack/placement-db-create-kf8dq" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.078133 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0e72c3-db89-4255-9c0b-50734039276b-operator-scripts\") pod \"placement-bb8e-account-create-update-fbh8q\" (UID: \"cf0e72c3-db89-4255-9c0b-50734039276b\") " pod="openstack/placement-bb8e-account-create-update-fbh8q" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.078546 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57db9b5bc9-6xt9l" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.093340 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w4jt\" (UniqueName: \"kubernetes.io/projected/cf0e72c3-db89-4255-9c0b-50734039276b-kube-api-access-4w4jt\") pod \"placement-bb8e-account-create-update-fbh8q\" (UID: \"cf0e72c3-db89-4255-9c0b-50734039276b\") " pod="openstack/placement-bb8e-account-create-update-fbh8q" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.102755 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmrfd\" (UniqueName: \"kubernetes.io/projected/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-kube-api-access-pmrfd\") pod \"placement-db-create-kf8dq\" (UID: \"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e\") " pod="openstack/placement-db-create-kf8dq" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.211558 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-6xt9l"] Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.216678 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57db9b5bc9-6xt9l"] Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.259828 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b445-account-create-update-84rcr" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.285050 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bb8e-account-create-update-fbh8q" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.297428 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kf8dq" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.444797 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92c86c07-aae3-45a0-b2fd-aa4532f2b542" path="/var/lib/kubelet/pods/92c86c07-aae3-45a0-b2fd-aa4532f2b542/volumes" Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.714425 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-9wd7r"] Dec 05 09:27:19 crc kubenswrapper[4815]: I1205 09:27:19.779932 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-b445-account-create-update-84rcr"] Dec 05 09:27:20 crc kubenswrapper[4815]: I1205 09:27:20.104231 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9wd7r" event={"ID":"bdc590e0-1c3a-418e-828b-eb43990319de","Type":"ContainerStarted","Data":"98d9f99ec2007edf338abda5debf396c8ab829f22682ae3edc797fac10fbb0c4"} Dec 05 09:27:20 crc kubenswrapper[4815]: I1205 09:27:20.109384 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b445-account-create-update-84rcr" event={"ID":"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd","Type":"ContainerStarted","Data":"9a00b31c05ae55f0e468476c5ef11c22a2a3b967fbb03fd38fdb750c3889ab69"} Dec 05 09:27:20 crc kubenswrapper[4815]: I1205 09:27:20.110940 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c3e63e92-1982-4055-9df7-db07096a570e","Type":"ContainerStarted","Data":"ef4afbb24f51971d38110209212e54cf2300b8a26557a545b7bafa45d671090f"} Dec 05 09:27:20 crc kubenswrapper[4815]: I1205 09:27:20.239289 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-kf8dq"] Dec 05 09:27:20 crc kubenswrapper[4815]: W1205 09:27:20.249264 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode31d5e84_0c31_4e29_a234_3d2ba4c45c3e.slice/crio-02dc9a609b681c2e452f39624d97a9ffe2c0c8c7fcd463bf0c90df73dc4dbee7 WatchSource:0}: Error finding container 02dc9a609b681c2e452f39624d97a9ffe2c0c8c7fcd463bf0c90df73dc4dbee7: Status 404 returned error can't find the container with id 02dc9a609b681c2e452f39624d97a9ffe2c0c8c7fcd463bf0c90df73dc4dbee7 Dec 05 09:27:20 crc kubenswrapper[4815]: I1205 09:27:20.324150 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-bb8e-account-create-update-fbh8q"] Dec 05 09:27:21 crc kubenswrapper[4815]: I1205 09:27:21.118518 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kf8dq" event={"ID":"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e","Type":"ContainerStarted","Data":"02dc9a609b681c2e452f39624d97a9ffe2c0c8c7fcd463bf0c90df73dc4dbee7"} Dec 05 09:27:21 crc kubenswrapper[4815]: I1205 09:27:21.119530 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bb8e-account-create-update-fbh8q" event={"ID":"cf0e72c3-db89-4255-9c0b-50734039276b","Type":"ContainerStarted","Data":"2ec7f7cb995841ce9401256e00a0469989b20d69e708500a51922d79ca02c246"} Dec 05 09:27:23 crc kubenswrapper[4815]: I1205 09:27:23.135766 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c3e63e92-1982-4055-9df7-db07096a570e","Type":"ContainerStarted","Data":"6c383c0c158b570c489d0da6d6f707ef5a9781e46bc2dcbf6c402f65179e8ab5"} Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.105043 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.151608 4815 generic.go:334] "Generic (PLEG): container finished" podID="bdc590e0-1c3a-418e-828b-eb43990319de" containerID="24e69df98780efdeaa978b1f4f10219cbb3516162678605a6c91f1fb00c6dd99" exitCode=0 Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.151708 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9wd7r" event={"ID":"bdc590e0-1c3a-418e-828b-eb43990319de","Type":"ContainerDied","Data":"24e69df98780efdeaa978b1f4f10219cbb3516162678605a6c91f1fb00c6dd99"} Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.153804 4815 generic.go:334] "Generic (PLEG): container finished" podID="7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd" containerID="6752875772ae7fecf27a3d608a95ec2fa45a5d6648ebcf06b347addc2ed51eb2" exitCode=0 Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.153877 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b445-account-create-update-84rcr" event={"ID":"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd","Type":"ContainerDied","Data":"6752875772ae7fecf27a3d608a95ec2fa45a5d6648ebcf06b347addc2ed51eb2"} Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.155989 4815 generic.go:334] "Generic (PLEG): container finished" podID="e31d5e84-0c31-4e29-a234-3d2ba4c45c3e" containerID="dbee94149eacb41f50efcef2b2fd7b41b04dc3f2a2b1bd52e79baccbc4bf0b1e" exitCode=0 Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.156058 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kf8dq" event={"ID":"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e","Type":"ContainerDied","Data":"dbee94149eacb41f50efcef2b2fd7b41b04dc3f2a2b1bd52e79baccbc4bf0b1e"} Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.158044 4815 generic.go:334] "Generic (PLEG): container finished" podID="cf0e72c3-db89-4255-9c0b-50734039276b" containerID="e3cb45ebf1c6ef6bfe6cf41744782bd99fea5008dd21d0fe01295c1ea2d23829" exitCode=0 Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.158110 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bb8e-account-create-update-fbh8q" event={"ID":"cf0e72c3-db89-4255-9c0b-50734039276b","Type":"ContainerDied","Data":"e3cb45ebf1c6ef6bfe6cf41744782bd99fea5008dd21d0fe01295c1ea2d23829"} Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.158265 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.258943 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=6.373792705 podStartE2EDuration="8.258924071s" podCreationTimestamp="2025-12-05 09:27:16 +0000 UTC" firstStartedPulling="2025-12-05 09:27:17.203173735 +0000 UTC m=+1236.081780572" lastFinishedPulling="2025-12-05 09:27:19.088305111 +0000 UTC m=+1237.966911938" observedRunningTime="2025-12-05 09:27:24.257449139 +0000 UTC m=+1243.136055986" watchObservedRunningTime="2025-12-05 09:27:24.258924071 +0000 UTC m=+1243.137530908" Dec 05 09:27:24 crc kubenswrapper[4815]: I1205 09:27:24.275682 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.552024 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bb8e-account-create-update-fbh8q" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.698061 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0e72c3-db89-4255-9c0b-50734039276b-operator-scripts\") pod \"cf0e72c3-db89-4255-9c0b-50734039276b\" (UID: \"cf0e72c3-db89-4255-9c0b-50734039276b\") " Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.698150 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4w4jt\" (UniqueName: \"kubernetes.io/projected/cf0e72c3-db89-4255-9c0b-50734039276b-kube-api-access-4w4jt\") pod \"cf0e72c3-db89-4255-9c0b-50734039276b\" (UID: \"cf0e72c3-db89-4255-9c0b-50734039276b\") " Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.700321 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf0e72c3-db89-4255-9c0b-50734039276b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cf0e72c3-db89-4255-9c0b-50734039276b" (UID: "cf0e72c3-db89-4255-9c0b-50734039276b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.712462 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b445-account-create-update-84rcr" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.717704 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf0e72c3-db89-4255-9c0b-50734039276b-kube-api-access-4w4jt" (OuterVolumeSpecName: "kube-api-access-4w4jt") pod "cf0e72c3-db89-4255-9c0b-50734039276b" (UID: "cf0e72c3-db89-4255-9c0b-50734039276b"). InnerVolumeSpecName "kube-api-access-4w4jt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.808182 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9wd7r" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.817826 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cf0e72c3-db89-4255-9c0b-50734039276b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.817865 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4w4jt\" (UniqueName: \"kubernetes.io/projected/cf0e72c3-db89-4255-9c0b-50734039276b-kube-api-access-4w4jt\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.837623 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kf8dq" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.928023 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs9f9\" (UniqueName: \"kubernetes.io/projected/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-kube-api-access-xs9f9\") pod \"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd\" (UID: \"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd\") " Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.928179 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-operator-scripts\") pod \"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd\" (UID: \"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd\") " Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.928212 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkzj2\" (UniqueName: \"kubernetes.io/projected/bdc590e0-1c3a-418e-828b-eb43990319de-kube-api-access-kkzj2\") pod \"bdc590e0-1c3a-418e-828b-eb43990319de\" (UID: \"bdc590e0-1c3a-418e-828b-eb43990319de\") " Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.928279 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-operator-scripts\") pod \"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e\" (UID: \"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e\") " Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.928310 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdc590e0-1c3a-418e-828b-eb43990319de-operator-scripts\") pod \"bdc590e0-1c3a-418e-828b-eb43990319de\" (UID: \"bdc590e0-1c3a-418e-828b-eb43990319de\") " Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.928375 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmrfd\" (UniqueName: \"kubernetes.io/projected/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-kube-api-access-pmrfd\") pod \"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e\" (UID: \"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e\") " Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.928787 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd" (UID: "7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.929195 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e31d5e84-0c31-4e29-a234-3d2ba4c45c3e" (UID: "e31d5e84-0c31-4e29-a234-3d2ba4c45c3e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.929242 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bdc590e0-1c3a-418e-828b-eb43990319de-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bdc590e0-1c3a-418e-828b-eb43990319de" (UID: "bdc590e0-1c3a-418e-828b-eb43990319de"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.932897 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-kube-api-access-xs9f9" (OuterVolumeSpecName: "kube-api-access-xs9f9") pod "7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd" (UID: "7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd"). InnerVolumeSpecName "kube-api-access-xs9f9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.937726 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bdc590e0-1c3a-418e-828b-eb43990319de-kube-api-access-kkzj2" (OuterVolumeSpecName: "kube-api-access-kkzj2") pod "bdc590e0-1c3a-418e-828b-eb43990319de" (UID: "bdc590e0-1c3a-418e-828b-eb43990319de"). InnerVolumeSpecName "kube-api-access-kkzj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:25 crc kubenswrapper[4815]: I1205 09:27:25.939895 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-kube-api-access-pmrfd" (OuterVolumeSpecName: "kube-api-access-pmrfd") pod "e31d5e84-0c31-4e29-a234-3d2ba4c45c3e" (UID: "e31d5e84-0c31-4e29-a234-3d2ba4c45c3e"). InnerVolumeSpecName "kube-api-access-pmrfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.030502 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs9f9\" (UniqueName: \"kubernetes.io/projected/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-kube-api-access-xs9f9\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.030545 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.030554 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkzj2\" (UniqueName: \"kubernetes.io/projected/bdc590e0-1c3a-418e-828b-eb43990319de-kube-api-access-kkzj2\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.030562 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.030571 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bdc590e0-1c3a-418e-828b-eb43990319de-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.030581 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmrfd\" (UniqueName: \"kubernetes.io/projected/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e-kube-api-access-pmrfd\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.171853 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-bb8e-account-create-update-fbh8q" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.171873 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-bb8e-account-create-update-fbh8q" event={"ID":"cf0e72c3-db89-4255-9c0b-50734039276b","Type":"ContainerDied","Data":"2ec7f7cb995841ce9401256e00a0469989b20d69e708500a51922d79ca02c246"} Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.171916 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ec7f7cb995841ce9401256e00a0469989b20d69e708500a51922d79ca02c246" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.173035 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-9wd7r" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.173051 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-9wd7r" event={"ID":"bdc590e0-1c3a-418e-828b-eb43990319de","Type":"ContainerDied","Data":"98d9f99ec2007edf338abda5debf396c8ab829f22682ae3edc797fac10fbb0c4"} Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.173085 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98d9f99ec2007edf338abda5debf396c8ab829f22682ae3edc797fac10fbb0c4" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.174122 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-b445-account-create-update-84rcr" event={"ID":"7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd","Type":"ContainerDied","Data":"9a00b31c05ae55f0e468476c5ef11c22a2a3b967fbb03fd38fdb750c3889ab69"} Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.174149 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a00b31c05ae55f0e468476c5ef11c22a2a3b967fbb03fd38fdb750c3889ab69" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.174188 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-b445-account-create-update-84rcr" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.179959 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-kf8dq" event={"ID":"e31d5e84-0c31-4e29-a234-3d2ba4c45c3e","Type":"ContainerDied","Data":"02dc9a609b681c2e452f39624d97a9ffe2c0c8c7fcd463bf0c90df73dc4dbee7"} Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.180100 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02dc9a609b681c2e452f39624d97a9ffe2c0c8c7fcd463bf0c90df73dc4dbee7" Dec 05 09:27:26 crc kubenswrapper[4815]: I1205 09:27:26.180043 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-kf8dq" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.069690 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-3451-account-create-update-thcm9"] Dec 05 09:27:28 crc kubenswrapper[4815]: E1205 09:27:28.070259 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd" containerName="mariadb-account-create-update" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.070285 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd" containerName="mariadb-account-create-update" Dec 05 09:27:28 crc kubenswrapper[4815]: E1205 09:27:28.070298 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf0e72c3-db89-4255-9c0b-50734039276b" containerName="mariadb-account-create-update" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.070304 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf0e72c3-db89-4255-9c0b-50734039276b" containerName="mariadb-account-create-update" Dec 05 09:27:28 crc kubenswrapper[4815]: E1205 09:27:28.070315 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e31d5e84-0c31-4e29-a234-3d2ba4c45c3e" containerName="mariadb-database-create" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.070322 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e31d5e84-0c31-4e29-a234-3d2ba4c45c3e" containerName="mariadb-database-create" Dec 05 09:27:28 crc kubenswrapper[4815]: E1205 09:27:28.070331 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bdc590e0-1c3a-418e-828b-eb43990319de" containerName="mariadb-database-create" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.070337 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="bdc590e0-1c3a-418e-828b-eb43990319de" containerName="mariadb-database-create" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.070521 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="bdc590e0-1c3a-418e-828b-eb43990319de" containerName="mariadb-database-create" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.070536 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf0e72c3-db89-4255-9c0b-50734039276b" containerName="mariadb-account-create-update" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.070548 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e31d5e84-0c31-4e29-a234-3d2ba4c45c3e" containerName="mariadb-database-create" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.070557 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd" containerName="mariadb-account-create-update" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.071133 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3451-account-create-update-thcm9" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.073346 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.078955 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-bjzq2"] Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.081571 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bjzq2" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.090110 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bjzq2"] Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.136220 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3451-account-create-update-thcm9"] Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.164784 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3231e31e-80a7-4aec-a346-2d34f455b62a-operator-scripts\") pod \"keystone-3451-account-create-update-thcm9\" (UID: \"3231e31e-80a7-4aec-a346-2d34f455b62a\") " pod="openstack/keystone-3451-account-create-update-thcm9" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.164847 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9170f00-6a0c-45db-86ca-906ed973f9bb-operator-scripts\") pod \"keystone-db-create-bjzq2\" (UID: \"f9170f00-6a0c-45db-86ca-906ed973f9bb\") " pod="openstack/keystone-db-create-bjzq2" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.164900 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lrd22\" (UniqueName: \"kubernetes.io/projected/3231e31e-80a7-4aec-a346-2d34f455b62a-kube-api-access-lrd22\") pod \"keystone-3451-account-create-update-thcm9\" (UID: \"3231e31e-80a7-4aec-a346-2d34f455b62a\") " pod="openstack/keystone-3451-account-create-update-thcm9" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.164961 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jf65\" (UniqueName: \"kubernetes.io/projected/f9170f00-6a0c-45db-86ca-906ed973f9bb-kube-api-access-5jf65\") pod \"keystone-db-create-bjzq2\" (UID: \"f9170f00-6a0c-45db-86ca-906ed973f9bb\") " pod="openstack/keystone-db-create-bjzq2" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.266635 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9170f00-6a0c-45db-86ca-906ed973f9bb-operator-scripts\") pod \"keystone-db-create-bjzq2\" (UID: \"f9170f00-6a0c-45db-86ca-906ed973f9bb\") " pod="openstack/keystone-db-create-bjzq2" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.266717 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lrd22\" (UniqueName: \"kubernetes.io/projected/3231e31e-80a7-4aec-a346-2d34f455b62a-kube-api-access-lrd22\") pod \"keystone-3451-account-create-update-thcm9\" (UID: \"3231e31e-80a7-4aec-a346-2d34f455b62a\") " pod="openstack/keystone-3451-account-create-update-thcm9" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.266799 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jf65\" (UniqueName: \"kubernetes.io/projected/f9170f00-6a0c-45db-86ca-906ed973f9bb-kube-api-access-5jf65\") pod \"keystone-db-create-bjzq2\" (UID: \"f9170f00-6a0c-45db-86ca-906ed973f9bb\") " pod="openstack/keystone-db-create-bjzq2" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.266974 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3231e31e-80a7-4aec-a346-2d34f455b62a-operator-scripts\") pod \"keystone-3451-account-create-update-thcm9\" (UID: \"3231e31e-80a7-4aec-a346-2d34f455b62a\") " pod="openstack/keystone-3451-account-create-update-thcm9" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.267647 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9170f00-6a0c-45db-86ca-906ed973f9bb-operator-scripts\") pod \"keystone-db-create-bjzq2\" (UID: \"f9170f00-6a0c-45db-86ca-906ed973f9bb\") " pod="openstack/keystone-db-create-bjzq2" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.267831 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3231e31e-80a7-4aec-a346-2d34f455b62a-operator-scripts\") pod \"keystone-3451-account-create-update-thcm9\" (UID: \"3231e31e-80a7-4aec-a346-2d34f455b62a\") " pod="openstack/keystone-3451-account-create-update-thcm9" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.285673 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lrd22\" (UniqueName: \"kubernetes.io/projected/3231e31e-80a7-4aec-a346-2d34f455b62a-kube-api-access-lrd22\") pod \"keystone-3451-account-create-update-thcm9\" (UID: \"3231e31e-80a7-4aec-a346-2d34f455b62a\") " pod="openstack/keystone-3451-account-create-update-thcm9" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.288978 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jf65\" (UniqueName: \"kubernetes.io/projected/f9170f00-6a0c-45db-86ca-906ed973f9bb-kube-api-access-5jf65\") pod \"keystone-db-create-bjzq2\" (UID: \"f9170f00-6a0c-45db-86ca-906ed973f9bb\") " pod="openstack/keystone-db-create-bjzq2" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.398375 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3451-account-create-update-thcm9" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.420394 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bjzq2" Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.871738 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-3451-account-create-update-thcm9"] Dec 05 09:27:28 crc kubenswrapper[4815]: I1205 09:27:28.959663 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bjzq2"] Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.089025 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-wjc6t"] Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.090081 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.093733 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-wng5w" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.101657 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-wjc6t"] Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.138897 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.195622 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-db-sync-config-data\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.195770 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-config-data\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.195810 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k46vx\" (UniqueName: \"kubernetes.io/projected/d00433e1-638a-4b77-bd8b-9052d9c9bc11-kube-api-access-k46vx\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.195849 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-combined-ca-bundle\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.242777 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3451-account-create-update-thcm9" event={"ID":"3231e31e-80a7-4aec-a346-2d34f455b62a","Type":"ContainerStarted","Data":"bae28e56dcceb530d59557ad2918bdf0be3088cb81734c3dde78d8a83654a904"} Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.242838 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3451-account-create-update-thcm9" event={"ID":"3231e31e-80a7-4aec-a346-2d34f455b62a","Type":"ContainerStarted","Data":"6d5baa41a91e60beeee023dab46d80485e786c3e5384da2e8e5c3fa1f3d1c0b3"} Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.266705 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bjzq2" event={"ID":"f9170f00-6a0c-45db-86ca-906ed973f9bb","Type":"ContainerStarted","Data":"00b401931da8f3f9670f76b993331352459ea936a193f066219d36fdc2aabcb1"} Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.267638 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-3451-account-create-update-thcm9" podStartSLOduration=1.2676193 podStartE2EDuration="1.2676193s" podCreationTimestamp="2025-12-05 09:27:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:27:29.266984842 +0000 UTC m=+1248.145591679" watchObservedRunningTime="2025-12-05 09:27:29.2676193 +0000 UTC m=+1248.146226137" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.297684 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-config-data\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.297759 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k46vx\" (UniqueName: \"kubernetes.io/projected/d00433e1-638a-4b77-bd8b-9052d9c9bc11-kube-api-access-k46vx\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.297777 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-combined-ca-bundle\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.297856 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-db-sync-config-data\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.308389 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-db-sync-config-data\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.309191 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-config-data\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.317672 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-combined-ca-bundle\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.354209 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k46vx\" (UniqueName: \"kubernetes.io/projected/d00433e1-638a-4b77-bd8b-9052d9c9bc11-kube-api-access-k46vx\") pod \"glance-db-sync-wjc6t\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:29 crc kubenswrapper[4815]: I1205 09:27:29.457249 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wjc6t" Dec 05 09:27:30 crc kubenswrapper[4815]: I1205 09:27:30.069911 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-wjc6t"] Dec 05 09:27:30 crc kubenswrapper[4815]: W1205 09:27:30.072422 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd00433e1_638a_4b77_bd8b_9052d9c9bc11.slice/crio-a8caca550a24792c339fb0c3c512a9e1842dbacda8abc64154575e41e25678b7 WatchSource:0}: Error finding container a8caca550a24792c339fb0c3c512a9e1842dbacda8abc64154575e41e25678b7: Status 404 returned error can't find the container with id a8caca550a24792c339fb0c3c512a9e1842dbacda8abc64154575e41e25678b7 Dec 05 09:27:30 crc kubenswrapper[4815]: I1205 09:27:30.274201 4815 generic.go:334] "Generic (PLEG): container finished" podID="3231e31e-80a7-4aec-a346-2d34f455b62a" containerID="bae28e56dcceb530d59557ad2918bdf0be3088cb81734c3dde78d8a83654a904" exitCode=0 Dec 05 09:27:30 crc kubenswrapper[4815]: I1205 09:27:30.274269 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3451-account-create-update-thcm9" event={"ID":"3231e31e-80a7-4aec-a346-2d34f455b62a","Type":"ContainerDied","Data":"bae28e56dcceb530d59557ad2918bdf0be3088cb81734c3dde78d8a83654a904"} Dec 05 09:27:30 crc kubenswrapper[4815]: I1205 09:27:30.277000 4815 generic.go:334] "Generic (PLEG): container finished" podID="f9170f00-6a0c-45db-86ca-906ed973f9bb" containerID="73609fb74300fe5e6c337116b71b5423ea6d50fd734c03b5193239a0834a646e" exitCode=0 Dec 05 09:27:30 crc kubenswrapper[4815]: I1205 09:27:30.277120 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bjzq2" event={"ID":"f9170f00-6a0c-45db-86ca-906ed973f9bb","Type":"ContainerDied","Data":"73609fb74300fe5e6c337116b71b5423ea6d50fd734c03b5193239a0834a646e"} Dec 05 09:27:30 crc kubenswrapper[4815]: I1205 09:27:30.278375 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wjc6t" event={"ID":"d00433e1-638a-4b77-bd8b-9052d9c9bc11","Type":"ContainerStarted","Data":"a8caca550a24792c339fb0c3c512a9e1842dbacda8abc64154575e41e25678b7"} Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.728383 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.816857 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3451-account-create-update-thcm9" Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.831425 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bjzq2" Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.856746 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lrd22\" (UniqueName: \"kubernetes.io/projected/3231e31e-80a7-4aec-a346-2d34f455b62a-kube-api-access-lrd22\") pod \"3231e31e-80a7-4aec-a346-2d34f455b62a\" (UID: \"3231e31e-80a7-4aec-a346-2d34f455b62a\") " Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.856854 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3231e31e-80a7-4aec-a346-2d34f455b62a-operator-scripts\") pod \"3231e31e-80a7-4aec-a346-2d34f455b62a\" (UID: \"3231e31e-80a7-4aec-a346-2d34f455b62a\") " Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.858378 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3231e31e-80a7-4aec-a346-2d34f455b62a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3231e31e-80a7-4aec-a346-2d34f455b62a" (UID: "3231e31e-80a7-4aec-a346-2d34f455b62a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.865992 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3231e31e-80a7-4aec-a346-2d34f455b62a-kube-api-access-lrd22" (OuterVolumeSpecName: "kube-api-access-lrd22") pod "3231e31e-80a7-4aec-a346-2d34f455b62a" (UID: "3231e31e-80a7-4aec-a346-2d34f455b62a"). InnerVolumeSpecName "kube-api-access-lrd22". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.959363 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jf65\" (UniqueName: \"kubernetes.io/projected/f9170f00-6a0c-45db-86ca-906ed973f9bb-kube-api-access-5jf65\") pod \"f9170f00-6a0c-45db-86ca-906ed973f9bb\" (UID: \"f9170f00-6a0c-45db-86ca-906ed973f9bb\") " Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.959571 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9170f00-6a0c-45db-86ca-906ed973f9bb-operator-scripts\") pod \"f9170f00-6a0c-45db-86ca-906ed973f9bb\" (UID: \"f9170f00-6a0c-45db-86ca-906ed973f9bb\") " Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.959996 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lrd22\" (UniqueName: \"kubernetes.io/projected/3231e31e-80a7-4aec-a346-2d34f455b62a-kube-api-access-lrd22\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.960018 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3231e31e-80a7-4aec-a346-2d34f455b62a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.960301 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9170f00-6a0c-45db-86ca-906ed973f9bb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f9170f00-6a0c-45db-86ca-906ed973f9bb" (UID: "f9170f00-6a0c-45db-86ca-906ed973f9bb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:31 crc kubenswrapper[4815]: I1205 09:27:31.965320 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9170f00-6a0c-45db-86ca-906ed973f9bb-kube-api-access-5jf65" (OuterVolumeSpecName: "kube-api-access-5jf65") pod "f9170f00-6a0c-45db-86ca-906ed973f9bb" (UID: "f9170f00-6a0c-45db-86ca-906ed973f9bb"). InnerVolumeSpecName "kube-api-access-5jf65". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:32 crc kubenswrapper[4815]: I1205 09:27:32.061404 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f9170f00-6a0c-45db-86ca-906ed973f9bb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:32 crc kubenswrapper[4815]: I1205 09:27:32.061447 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jf65\" (UniqueName: \"kubernetes.io/projected/f9170f00-6a0c-45db-86ca-906ed973f9bb-kube-api-access-5jf65\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:32 crc kubenswrapper[4815]: I1205 09:27:32.363434 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-3451-account-create-update-thcm9" event={"ID":"3231e31e-80a7-4aec-a346-2d34f455b62a","Type":"ContainerDied","Data":"6d5baa41a91e60beeee023dab46d80485e786c3e5384da2e8e5c3fa1f3d1c0b3"} Dec 05 09:27:32 crc kubenswrapper[4815]: I1205 09:27:32.363475 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d5baa41a91e60beeee023dab46d80485e786c3e5384da2e8e5c3fa1f3d1c0b3" Dec 05 09:27:32 crc kubenswrapper[4815]: I1205 09:27:32.363477 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-3451-account-create-update-thcm9" Dec 05 09:27:32 crc kubenswrapper[4815]: I1205 09:27:32.369841 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bjzq2" event={"ID":"f9170f00-6a0c-45db-86ca-906ed973f9bb","Type":"ContainerDied","Data":"00b401931da8f3f9670f76b993331352459ea936a193f066219d36fdc2aabcb1"} Dec 05 09:27:32 crc kubenswrapper[4815]: I1205 09:27:32.369878 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="00b401931da8f3f9670f76b993331352459ea936a193f066219d36fdc2aabcb1" Dec 05 09:27:32 crc kubenswrapper[4815]: I1205 09:27:32.369938 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bjzq2" Dec 05 09:27:35 crc kubenswrapper[4815]: I1205 09:27:35.392805 4815 generic.go:334] "Generic (PLEG): container finished" podID="e965503d-d60a-4ce1-9dee-b74fd5eb66a0" containerID="396633accd6c07faf0270b697296080712b2c885fe98668cd948cdf27a9bfbfa" exitCode=0 Dec 05 09:27:35 crc kubenswrapper[4815]: I1205 09:27:35.392893 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e965503d-d60a-4ce1-9dee-b74fd5eb66a0","Type":"ContainerDied","Data":"396633accd6c07faf0270b697296080712b2c885fe98668cd948cdf27a9bfbfa"} Dec 05 09:27:35 crc kubenswrapper[4815]: I1205 09:27:35.396883 4815 generic.go:334] "Generic (PLEG): container finished" podID="852e4372-ed4f-4350-b013-ab36741cd980" containerID="fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186" exitCode=0 Dec 05 09:27:35 crc kubenswrapper[4815]: I1205 09:27:35.396929 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"852e4372-ed4f-4350-b013-ab36741cd980","Type":"ContainerDied","Data":"fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186"} Dec 05 09:27:36 crc kubenswrapper[4815]: I1205 09:27:36.406808 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e965503d-d60a-4ce1-9dee-b74fd5eb66a0","Type":"ContainerStarted","Data":"175e366dd5e953a14c204f9dec873e4e240ebdd20e12c52e0d7b74e0d64573cb"} Dec 05 09:27:36 crc kubenswrapper[4815]: I1205 09:27:36.407257 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 05 09:27:36 crc kubenswrapper[4815]: I1205 09:27:36.408971 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"852e4372-ed4f-4350-b013-ab36741cd980","Type":"ContainerStarted","Data":"c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4"} Dec 05 09:27:36 crc kubenswrapper[4815]: I1205 09:27:36.409185 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:27:36 crc kubenswrapper[4815]: I1205 09:27:36.436831 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=39.723570421 podStartE2EDuration="1m33.436791853s" podCreationTimestamp="2025-12-05 09:26:03 +0000 UTC" firstStartedPulling="2025-12-05 09:26:05.562184274 +0000 UTC m=+1164.440791111" lastFinishedPulling="2025-12-05 09:26:59.275405706 +0000 UTC m=+1218.154012543" observedRunningTime="2025-12-05 09:27:36.430542392 +0000 UTC m=+1255.309149239" watchObservedRunningTime="2025-12-05 09:27:36.436791853 +0000 UTC m=+1255.315398690" Dec 05 09:27:36 crc kubenswrapper[4815]: I1205 09:27:36.465560 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.520413837 podStartE2EDuration="1m33.465533257s" podCreationTimestamp="2025-12-05 09:26:03 +0000 UTC" firstStartedPulling="2025-12-05 09:26:05.864226645 +0000 UTC m=+1164.742833482" lastFinishedPulling="2025-12-05 09:27:00.809346075 +0000 UTC m=+1219.687952902" observedRunningTime="2025-12-05 09:27:36.457198685 +0000 UTC m=+1255.335805542" watchObservedRunningTime="2025-12-05 09:27:36.465533257 +0000 UTC m=+1255.344140094" Dec 05 09:27:43 crc kubenswrapper[4815]: I1205 09:27:43.690527 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-v5mcb" podUID="0e0fc690-fb2a-4381-93df-1adcb21c0c7d" containerName="ovn-controller" probeResult="failure" output=< Dec 05 09:27:43 crc kubenswrapper[4815]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 05 09:27:43 crc kubenswrapper[4815]: > Dec 05 09:27:43 crc kubenswrapper[4815]: I1205 09:27:43.917450 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.010342 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zlv59" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.345775 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-v5mcb-config-2d88s"] Dec 05 09:27:44 crc kubenswrapper[4815]: E1205 09:27:44.346181 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3231e31e-80a7-4aec-a346-2d34f455b62a" containerName="mariadb-account-create-update" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.346213 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="3231e31e-80a7-4aec-a346-2d34f455b62a" containerName="mariadb-account-create-update" Dec 05 09:27:44 crc kubenswrapper[4815]: E1205 09:27:44.346262 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9170f00-6a0c-45db-86ca-906ed973f9bb" containerName="mariadb-database-create" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.346272 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9170f00-6a0c-45db-86ca-906ed973f9bb" containerName="mariadb-database-create" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.346526 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="3231e31e-80a7-4aec-a346-2d34f455b62a" containerName="mariadb-account-create-update" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.346556 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9170f00-6a0c-45db-86ca-906ed973f9bb" containerName="mariadb-database-create" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.347204 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.349390 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.381159 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-v5mcb-config-2d88s"] Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.563356 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwdvr\" (UniqueName: \"kubernetes.io/projected/f2910008-2b8d-4faa-b721-c50b0fd3eb40-kube-api-access-lwdvr\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.563845 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run-ovn\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.563913 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-log-ovn\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.563943 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-scripts\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.564007 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-additional-scripts\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.564055 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.665846 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.665902 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwdvr\" (UniqueName: \"kubernetes.io/projected/f2910008-2b8d-4faa-b721-c50b0fd3eb40-kube-api-access-lwdvr\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.665976 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run-ovn\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.665996 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-log-ovn\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.666013 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-scripts\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.666040 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-additional-scripts\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.666130 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.666187 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-log-ovn\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.666213 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run-ovn\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.666725 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-additional-scripts\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.667845 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-scripts\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.710305 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwdvr\" (UniqueName: \"kubernetes.io/projected/f2910008-2b8d-4faa-b721-c50b0fd3eb40-kube-api-access-lwdvr\") pod \"ovn-controller-v5mcb-config-2d88s\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:44 crc kubenswrapper[4815]: I1205 09:27:44.973182 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:45 crc kubenswrapper[4815]: I1205 09:27:45.316390 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="852e4372-ed4f-4350-b013-ab36741cd980" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 05 09:27:47 crc kubenswrapper[4815]: E1205 09:27:47.356855 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63" Dec 05 09:27:47 crc kubenswrapper[4815]: E1205 09:27:47.357295 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k46vx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-wjc6t_openstack(d00433e1-638a-4b77-bd8b-9052d9c9bc11): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:27:47 crc kubenswrapper[4815]: E1205 09:27:47.358481 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-wjc6t" podUID="d00433e1-638a-4b77-bd8b-9052d9c9bc11" Dec 05 09:27:47 crc kubenswrapper[4815]: E1205 09:27:47.501947 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api@sha256:481073ac9deefb38bbd989aaa8dd7aedb4e0af26017f4883f85fce433380bf63\\\"\"" pod="openstack/glance-db-sync-wjc6t" podUID="d00433e1-638a-4b77-bd8b-9052d9c9bc11" Dec 05 09:27:47 crc kubenswrapper[4815]: I1205 09:27:47.839928 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-v5mcb-config-2d88s"] Dec 05 09:27:48 crc kubenswrapper[4815]: I1205 09:27:48.526305 4815 generic.go:334] "Generic (PLEG): container finished" podID="f2910008-2b8d-4faa-b721-c50b0fd3eb40" containerID="aaf49eeb4c0b5fa43b82eaae0c60cc6675ab0c58985d81f0bf994f967b6863d4" exitCode=0 Dec 05 09:27:48 crc kubenswrapper[4815]: I1205 09:27:48.526419 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v5mcb-config-2d88s" event={"ID":"f2910008-2b8d-4faa-b721-c50b0fd3eb40","Type":"ContainerDied","Data":"aaf49eeb4c0b5fa43b82eaae0c60cc6675ab0c58985d81f0bf994f967b6863d4"} Dec 05 09:27:48 crc kubenswrapper[4815]: I1205 09:27:48.526583 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v5mcb-config-2d88s" event={"ID":"f2910008-2b8d-4faa-b721-c50b0fd3eb40","Type":"ContainerStarted","Data":"3ce2d1d30f5603309d19f940763abdb341beaf10e0c2e93ef5e250ad293f656a"} Dec 05 09:27:48 crc kubenswrapper[4815]: I1205 09:27:48.706545 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-v5mcb" Dec 05 09:27:49 crc kubenswrapper[4815]: I1205 09:27:49.874451 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.001636 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run-ovn\") pod \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.001694 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run\") pod \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.001743 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwdvr\" (UniqueName: \"kubernetes.io/projected/f2910008-2b8d-4faa-b721-c50b0fd3eb40-kube-api-access-lwdvr\") pod \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.001750 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f2910008-2b8d-4faa-b721-c50b0fd3eb40" (UID: "f2910008-2b8d-4faa-b721-c50b0fd3eb40"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.001789 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-additional-scripts\") pod \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.001792 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run" (OuterVolumeSpecName: "var-run") pod "f2910008-2b8d-4faa-b721-c50b0fd3eb40" (UID: "f2910008-2b8d-4faa-b721-c50b0fd3eb40"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.001826 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-scripts\") pod \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.001844 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-log-ovn\") pod \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\" (UID: \"f2910008-2b8d-4faa-b721-c50b0fd3eb40\") " Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.002200 4815 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.002216 4815 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-run\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.002260 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f2910008-2b8d-4faa-b721-c50b0fd3eb40" (UID: "f2910008-2b8d-4faa-b721-c50b0fd3eb40"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.002627 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f2910008-2b8d-4faa-b721-c50b0fd3eb40" (UID: "f2910008-2b8d-4faa-b721-c50b0fd3eb40"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.002962 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-scripts" (OuterVolumeSpecName: "scripts") pod "f2910008-2b8d-4faa-b721-c50b0fd3eb40" (UID: "f2910008-2b8d-4faa-b721-c50b0fd3eb40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.009852 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2910008-2b8d-4faa-b721-c50b0fd3eb40-kube-api-access-lwdvr" (OuterVolumeSpecName: "kube-api-access-lwdvr") pod "f2910008-2b8d-4faa-b721-c50b0fd3eb40" (UID: "f2910008-2b8d-4faa-b721-c50b0fd3eb40"). InnerVolumeSpecName "kube-api-access-lwdvr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.103391 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwdvr\" (UniqueName: \"kubernetes.io/projected/f2910008-2b8d-4faa-b721-c50b0fd3eb40-kube-api-access-lwdvr\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.103421 4815 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.103433 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f2910008-2b8d-4faa-b721-c50b0fd3eb40-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.103442 4815 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f2910008-2b8d-4faa-b721-c50b0fd3eb40-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.543849 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-v5mcb-config-2d88s" event={"ID":"f2910008-2b8d-4faa-b721-c50b0fd3eb40","Type":"ContainerDied","Data":"3ce2d1d30f5603309d19f940763abdb341beaf10e0c2e93ef5e250ad293f656a"} Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.544216 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ce2d1d30f5603309d19f940763abdb341beaf10e0c2e93ef5e250ad293f656a" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.543912 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-v5mcb-config-2d88s" Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.980031 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-v5mcb-config-2d88s"] Dec 05 09:27:50 crc kubenswrapper[4815]: I1205 09:27:50.986701 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-v5mcb-config-2d88s"] Dec 05 09:27:51 crc kubenswrapper[4815]: I1205 09:27:51.428898 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2910008-2b8d-4faa-b721-c50b0fd3eb40" path="/var/lib/kubelet/pods/f2910008-2b8d-4faa-b721-c50b0fd3eb40/volumes" Dec 05 09:27:54 crc kubenswrapper[4815]: I1205 09:27:54.942646 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.315638 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.668281 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-63a3-account-create-update-5hprm"] Dec 05 09:27:55 crc kubenswrapper[4815]: E1205 09:27:55.668732 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2910008-2b8d-4faa-b721-c50b0fd3eb40" containerName="ovn-config" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.668753 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2910008-2b8d-4faa-b721-c50b0fd3eb40" containerName="ovn-config" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.668953 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2910008-2b8d-4faa-b721-c50b0fd3eb40" containerName="ovn-config" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.669519 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-63a3-account-create-update-5hprm" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.671442 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.677145 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-z6t49"] Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.678115 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z6t49" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.706591 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-z6t49"] Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.714020 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-63a3-account-create-update-5hprm"] Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.812123 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dfe6e19-f7c0-4360-945c-e70be5e60932-operator-scripts\") pod \"barbican-63a3-account-create-update-5hprm\" (UID: \"6dfe6e19-f7c0-4360-945c-e70be5e60932\") " pod="openstack/barbican-63a3-account-create-update-5hprm" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.812205 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nkvqv\" (UniqueName: \"kubernetes.io/projected/6dfe6e19-f7c0-4360-945c-e70be5e60932-kube-api-access-nkvqv\") pod \"barbican-63a3-account-create-update-5hprm\" (UID: \"6dfe6e19-f7c0-4360-945c-e70be5e60932\") " pod="openstack/barbican-63a3-account-create-update-5hprm" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.812229 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45462c49-fb2a-48e8-bd00-98efa3a73da7-operator-scripts\") pod \"barbican-db-create-z6t49\" (UID: \"45462c49-fb2a-48e8-bd00-98efa3a73da7\") " pod="openstack/barbican-db-create-z6t49" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.812267 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk2v5\" (UniqueName: \"kubernetes.io/projected/45462c49-fb2a-48e8-bd00-98efa3a73da7-kube-api-access-fk2v5\") pod \"barbican-db-create-z6t49\" (UID: \"45462c49-fb2a-48e8-bd00-98efa3a73da7\") " pod="openstack/barbican-db-create-z6t49" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.826114 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-6pbnl"] Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.828642 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6pbnl" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.832820 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6pbnl"] Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.872313 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-d057-account-create-update-kdmmk"] Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.873279 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d057-account-create-update-kdmmk" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.875576 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.890801 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d057-account-create-update-kdmmk"] Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.922381 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dfe6e19-f7c0-4360-945c-e70be5e60932-operator-scripts\") pod \"barbican-63a3-account-create-update-5hprm\" (UID: \"6dfe6e19-f7c0-4360-945c-e70be5e60932\") " pod="openstack/barbican-63a3-account-create-update-5hprm" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.922437 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nkvqv\" (UniqueName: \"kubernetes.io/projected/6dfe6e19-f7c0-4360-945c-e70be5e60932-kube-api-access-nkvqv\") pod \"barbican-63a3-account-create-update-5hprm\" (UID: \"6dfe6e19-f7c0-4360-945c-e70be5e60932\") " pod="openstack/barbican-63a3-account-create-update-5hprm" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.922460 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45462c49-fb2a-48e8-bd00-98efa3a73da7-operator-scripts\") pod \"barbican-db-create-z6t49\" (UID: \"45462c49-fb2a-48e8-bd00-98efa3a73da7\") " pod="openstack/barbican-db-create-z6t49" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.922526 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk2v5\" (UniqueName: \"kubernetes.io/projected/45462c49-fb2a-48e8-bd00-98efa3a73da7-kube-api-access-fk2v5\") pod \"barbican-db-create-z6t49\" (UID: \"45462c49-fb2a-48e8-bd00-98efa3a73da7\") " pod="openstack/barbican-db-create-z6t49" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.922566 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vnvs\" (UniqueName: \"kubernetes.io/projected/b7a77236-882a-4a90-99bc-b74679d44e01-kube-api-access-6vnvs\") pod \"cinder-db-create-6pbnl\" (UID: \"b7a77236-882a-4a90-99bc-b74679d44e01\") " pod="openstack/cinder-db-create-6pbnl" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.922644 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7a77236-882a-4a90-99bc-b74679d44e01-operator-scripts\") pod \"cinder-db-create-6pbnl\" (UID: \"b7a77236-882a-4a90-99bc-b74679d44e01\") " pod="openstack/cinder-db-create-6pbnl" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.923384 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dfe6e19-f7c0-4360-945c-e70be5e60932-operator-scripts\") pod \"barbican-63a3-account-create-update-5hprm\" (UID: \"6dfe6e19-f7c0-4360-945c-e70be5e60932\") " pod="openstack/barbican-63a3-account-create-update-5hprm" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.923403 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45462c49-fb2a-48e8-bd00-98efa3a73da7-operator-scripts\") pod \"barbican-db-create-z6t49\" (UID: \"45462c49-fb2a-48e8-bd00-98efa3a73da7\") " pod="openstack/barbican-db-create-z6t49" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.939939 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nkvqv\" (UniqueName: \"kubernetes.io/projected/6dfe6e19-f7c0-4360-945c-e70be5e60932-kube-api-access-nkvqv\") pod \"barbican-63a3-account-create-update-5hprm\" (UID: \"6dfe6e19-f7c0-4360-945c-e70be5e60932\") " pod="openstack/barbican-63a3-account-create-update-5hprm" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.954909 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-2qpg9"] Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.956336 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2qpg9" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.970318 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk2v5\" (UniqueName: \"kubernetes.io/projected/45462c49-fb2a-48e8-bd00-98efa3a73da7-kube-api-access-fk2v5\") pod \"barbican-db-create-z6t49\" (UID: \"45462c49-fb2a-48e8-bd00-98efa3a73da7\") " pod="openstack/barbican-db-create-z6t49" Dec 05 09:27:55 crc kubenswrapper[4815]: I1205 09:27:55.978261 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2qpg9"] Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.002002 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z6t49" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.002143 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-63a3-account-create-update-5hprm" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.024637 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d044ebe-1e9f-4412-a94f-702501190349-operator-scripts\") pod \"cinder-d057-account-create-update-kdmmk\" (UID: \"9d044ebe-1e9f-4412-a94f-702501190349\") " pod="openstack/cinder-d057-account-create-update-kdmmk" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.024978 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vnvs\" (UniqueName: \"kubernetes.io/projected/b7a77236-882a-4a90-99bc-b74679d44e01-kube-api-access-6vnvs\") pod \"cinder-db-create-6pbnl\" (UID: \"b7a77236-882a-4a90-99bc-b74679d44e01\") " pod="openstack/cinder-db-create-6pbnl" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.025182 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7a77236-882a-4a90-99bc-b74679d44e01-operator-scripts\") pod \"cinder-db-create-6pbnl\" (UID: \"b7a77236-882a-4a90-99bc-b74679d44e01\") " pod="openstack/cinder-db-create-6pbnl" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.025324 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99gjr\" (UniqueName: \"kubernetes.io/projected/9d044ebe-1e9f-4412-a94f-702501190349-kube-api-access-99gjr\") pod \"cinder-d057-account-create-update-kdmmk\" (UID: \"9d044ebe-1e9f-4412-a94f-702501190349\") " pod="openstack/cinder-d057-account-create-update-kdmmk" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.025849 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7a77236-882a-4a90-99bc-b74679d44e01-operator-scripts\") pod \"cinder-db-create-6pbnl\" (UID: \"b7a77236-882a-4a90-99bc-b74679d44e01\") " pod="openstack/cinder-db-create-6pbnl" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.043540 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vnvs\" (UniqueName: \"kubernetes.io/projected/b7a77236-882a-4a90-99bc-b74679d44e01-kube-api-access-6vnvs\") pod \"cinder-db-create-6pbnl\" (UID: \"b7a77236-882a-4a90-99bc-b74679d44e01\") " pod="openstack/cinder-db-create-6pbnl" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.138325 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99gjr\" (UniqueName: \"kubernetes.io/projected/9d044ebe-1e9f-4412-a94f-702501190349-kube-api-access-99gjr\") pod \"cinder-d057-account-create-update-kdmmk\" (UID: \"9d044ebe-1e9f-4412-a94f-702501190349\") " pod="openstack/cinder-d057-account-create-update-kdmmk" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.138374 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d044ebe-1e9f-4412-a94f-702501190349-operator-scripts\") pod \"cinder-d057-account-create-update-kdmmk\" (UID: \"9d044ebe-1e9f-4412-a94f-702501190349\") " pod="openstack/cinder-d057-account-create-update-kdmmk" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.138407 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cn7fg\" (UniqueName: \"kubernetes.io/projected/2a6fff39-63be-4b79-a9a0-dbcf8162a231-kube-api-access-cn7fg\") pod \"neutron-db-create-2qpg9\" (UID: \"2a6fff39-63be-4b79-a9a0-dbcf8162a231\") " pod="openstack/neutron-db-create-2qpg9" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.138436 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a6fff39-63be-4b79-a9a0-dbcf8162a231-operator-scripts\") pod \"neutron-db-create-2qpg9\" (UID: \"2a6fff39-63be-4b79-a9a0-dbcf8162a231\") " pod="openstack/neutron-db-create-2qpg9" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.146619 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6pbnl" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.194044 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99gjr\" (UniqueName: \"kubernetes.io/projected/9d044ebe-1e9f-4412-a94f-702501190349-kube-api-access-99gjr\") pod \"cinder-d057-account-create-update-kdmmk\" (UID: \"9d044ebe-1e9f-4412-a94f-702501190349\") " pod="openstack/cinder-d057-account-create-update-kdmmk" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.240339 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cn7fg\" (UniqueName: \"kubernetes.io/projected/2a6fff39-63be-4b79-a9a0-dbcf8162a231-kube-api-access-cn7fg\") pod \"neutron-db-create-2qpg9\" (UID: \"2a6fff39-63be-4b79-a9a0-dbcf8162a231\") " pod="openstack/neutron-db-create-2qpg9" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.240394 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a6fff39-63be-4b79-a9a0-dbcf8162a231-operator-scripts\") pod \"neutron-db-create-2qpg9\" (UID: \"2a6fff39-63be-4b79-a9a0-dbcf8162a231\") " pod="openstack/neutron-db-create-2qpg9" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.247880 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a6fff39-63be-4b79-a9a0-dbcf8162a231-operator-scripts\") pod \"neutron-db-create-2qpg9\" (UID: \"2a6fff39-63be-4b79-a9a0-dbcf8162a231\") " pod="openstack/neutron-db-create-2qpg9" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.265230 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cn7fg\" (UniqueName: \"kubernetes.io/projected/2a6fff39-63be-4b79-a9a0-dbcf8162a231-kube-api-access-cn7fg\") pod \"neutron-db-create-2qpg9\" (UID: \"2a6fff39-63be-4b79-a9a0-dbcf8162a231\") " pod="openstack/neutron-db-create-2qpg9" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.331124 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d044ebe-1e9f-4412-a94f-702501190349-operator-scripts\") pod \"cinder-d057-account-create-update-kdmmk\" (UID: \"9d044ebe-1e9f-4412-a94f-702501190349\") " pod="openstack/cinder-d057-account-create-update-kdmmk" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.448856 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2qpg9" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.453929 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-z6t49"] Dec 05 09:27:56 crc kubenswrapper[4815]: W1205 09:27:56.485725 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45462c49_fb2a_48e8_bd00_98efa3a73da7.slice/crio-819bda1181cad3f6c7fe4c40d6c0f825ae4014f793ecd43ad41fcaa9662b16c2 WatchSource:0}: Error finding container 819bda1181cad3f6c7fe4c40d6c0f825ae4014f793ecd43ad41fcaa9662b16c2: Status 404 returned error can't find the container with id 819bda1181cad3f6c7fe4c40d6c0f825ae4014f793ecd43ad41fcaa9662b16c2 Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.497750 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d057-account-create-update-kdmmk" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.534059 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-63a3-account-create-update-5hprm"] Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.582014 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-c640-account-create-update-pb7l7"] Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.582987 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c640-account-create-update-pb7l7" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.587051 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.624195 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c640-account-create-update-pb7l7"] Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.630024 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-z6t49" event={"ID":"45462c49-fb2a-48e8-bd00-98efa3a73da7","Type":"ContainerStarted","Data":"819bda1181cad3f6c7fe4c40d6c0f825ae4014f793ecd43ad41fcaa9662b16c2"} Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.666688 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-bmkjh"] Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.667963 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.682552 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.682808 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xn54c" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.683393 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.683473 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.684392 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8373b63-0203-47e7-b23c-714afe735109-operator-scripts\") pod \"neutron-c640-account-create-update-pb7l7\" (UID: \"f8373b63-0203-47e7-b23c-714afe735109\") " pod="openstack/neutron-c640-account-create-update-pb7l7" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.684443 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-combined-ca-bundle\") pod \"keystone-db-sync-bmkjh\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.684465 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wfqpb\" (UniqueName: \"kubernetes.io/projected/f8373b63-0203-47e7-b23c-714afe735109-kube-api-access-wfqpb\") pod \"neutron-c640-account-create-update-pb7l7\" (UID: \"f8373b63-0203-47e7-b23c-714afe735109\") " pod="openstack/neutron-c640-account-create-update-pb7l7" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.684539 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntcd2\" (UniqueName: \"kubernetes.io/projected/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-kube-api-access-ntcd2\") pod \"keystone-db-sync-bmkjh\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.684599 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-config-data\") pod \"keystone-db-sync-bmkjh\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.725403 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bmkjh"] Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.788895 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-config-data\") pod \"keystone-db-sync-bmkjh\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.789098 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8373b63-0203-47e7-b23c-714afe735109-operator-scripts\") pod \"neutron-c640-account-create-update-pb7l7\" (UID: \"f8373b63-0203-47e7-b23c-714afe735109\") " pod="openstack/neutron-c640-account-create-update-pb7l7" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.789211 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-combined-ca-bundle\") pod \"keystone-db-sync-bmkjh\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.789293 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wfqpb\" (UniqueName: \"kubernetes.io/projected/f8373b63-0203-47e7-b23c-714afe735109-kube-api-access-wfqpb\") pod \"neutron-c640-account-create-update-pb7l7\" (UID: \"f8373b63-0203-47e7-b23c-714afe735109\") " pod="openstack/neutron-c640-account-create-update-pb7l7" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.789370 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntcd2\" (UniqueName: \"kubernetes.io/projected/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-kube-api-access-ntcd2\") pod \"keystone-db-sync-bmkjh\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.793271 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8373b63-0203-47e7-b23c-714afe735109-operator-scripts\") pod \"neutron-c640-account-create-update-pb7l7\" (UID: \"f8373b63-0203-47e7-b23c-714afe735109\") " pod="openstack/neutron-c640-account-create-update-pb7l7" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.939569 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-combined-ca-bundle\") pod \"keystone-db-sync-bmkjh\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.940089 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-config-data\") pod \"keystone-db-sync-bmkjh\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.972909 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wfqpb\" (UniqueName: \"kubernetes.io/projected/f8373b63-0203-47e7-b23c-714afe735109-kube-api-access-wfqpb\") pod \"neutron-c640-account-create-update-pb7l7\" (UID: \"f8373b63-0203-47e7-b23c-714afe735109\") " pod="openstack/neutron-c640-account-create-update-pb7l7" Dec 05 09:27:56 crc kubenswrapper[4815]: I1205 09:27:56.995249 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntcd2\" (UniqueName: \"kubernetes.io/projected/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-kube-api-access-ntcd2\") pod \"keystone-db-sync-bmkjh\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.044409 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.065033 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-6pbnl"] Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.225115 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c640-account-create-update-pb7l7" Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.493209 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d057-account-create-update-kdmmk"] Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.613316 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-2qpg9"] Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.708737 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-z6t49" event={"ID":"45462c49-fb2a-48e8-bd00-98efa3a73da7","Type":"ContainerStarted","Data":"7dfe05922f1b2c3eac292e7c7d589817b18eba9a50d15a97ad9db4e6beb2f1b2"} Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.753528 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-63a3-account-create-update-5hprm" event={"ID":"6dfe6e19-f7c0-4360-945c-e70be5e60932","Type":"ContainerStarted","Data":"5c28b7227af542c58bec1f7fc1608cd3cf22ddc730b8a6b3ecc9a1eeed7497ee"} Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.761150 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2qpg9" event={"ID":"2a6fff39-63be-4b79-a9a0-dbcf8162a231","Type":"ContainerStarted","Data":"396344cdb6bd79e66d62ec40c5512e37a48dce43809a9933acf89473387d2ed0"} Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.762562 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d057-account-create-update-kdmmk" event={"ID":"9d044ebe-1e9f-4412-a94f-702501190349","Type":"ContainerStarted","Data":"d1778a63e1862a7bf4b89d4601d517ec43342b0fd02c0f3817f663089dae7f3a"} Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.763731 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6pbnl" event={"ID":"b7a77236-882a-4a90-99bc-b74679d44e01","Type":"ContainerStarted","Data":"557ce497566d91e66527a2f6e3a825f6b29de51b692d2f5549058011f73a83db"} Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.767734 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-z6t49" podStartSLOduration=2.767709627 podStartE2EDuration="2.767709627s" podCreationTimestamp="2025-12-05 09:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:27:57.760368704 +0000 UTC m=+1276.638975541" watchObservedRunningTime="2025-12-05 09:27:57.767709627 +0000 UTC m=+1276.646316484" Dec 05 09:27:57 crc kubenswrapper[4815]: I1205 09:27:57.800756 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-6pbnl" podStartSLOduration=2.800735514 podStartE2EDuration="2.800735514s" podCreationTimestamp="2025-12-05 09:27:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:27:57.788733576 +0000 UTC m=+1276.667340413" watchObservedRunningTime="2025-12-05 09:27:57.800735514 +0000 UTC m=+1276.679342351" Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.071691 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-c640-account-create-update-pb7l7"] Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.177351 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-bmkjh"] Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.772956 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bmkjh" event={"ID":"a86ad8ca-5199-41c4-ab4e-22826ea4b65d","Type":"ContainerStarted","Data":"e7720da1d6348fe94047b9c8161280b90ecb01be3e8b7c29ed6e01dd0cea7293"} Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.775130 4815 generic.go:334] "Generic (PLEG): container finished" podID="b7a77236-882a-4a90-99bc-b74679d44e01" containerID="1b7e1ff9d5277936ebb7ec3ecfdeca681e7b2c8578088537eca5c2cde35f048f" exitCode=0 Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.775231 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6pbnl" event={"ID":"b7a77236-882a-4a90-99bc-b74679d44e01","Type":"ContainerDied","Data":"1b7e1ff9d5277936ebb7ec3ecfdeca681e7b2c8578088537eca5c2cde35f048f"} Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.777097 4815 generic.go:334] "Generic (PLEG): container finished" podID="45462c49-fb2a-48e8-bd00-98efa3a73da7" containerID="7dfe05922f1b2c3eac292e7c7d589817b18eba9a50d15a97ad9db4e6beb2f1b2" exitCode=0 Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.777175 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-z6t49" event={"ID":"45462c49-fb2a-48e8-bd00-98efa3a73da7","Type":"ContainerDied","Data":"7dfe05922f1b2c3eac292e7c7d589817b18eba9a50d15a97ad9db4e6beb2f1b2"} Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.778637 4815 generic.go:334] "Generic (PLEG): container finished" podID="6dfe6e19-f7c0-4360-945c-e70be5e60932" containerID="dda28bff0f5484c532faafe0684bae0d1d724980f07b9c53bff78ad1fe077a11" exitCode=0 Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.778679 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-63a3-account-create-update-5hprm" event={"ID":"6dfe6e19-f7c0-4360-945c-e70be5e60932","Type":"ContainerDied","Data":"dda28bff0f5484c532faafe0684bae0d1d724980f07b9c53bff78ad1fe077a11"} Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.781900 4815 generic.go:334] "Generic (PLEG): container finished" podID="2a6fff39-63be-4b79-a9a0-dbcf8162a231" containerID="41337d3bda688a2910f9311369522d3ffcf27c4c3d69ed3bce373f33009afaba" exitCode=0 Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.782043 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2qpg9" event={"ID":"2a6fff39-63be-4b79-a9a0-dbcf8162a231","Type":"ContainerDied","Data":"41337d3bda688a2910f9311369522d3ffcf27c4c3d69ed3bce373f33009afaba"} Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.784235 4815 generic.go:334] "Generic (PLEG): container finished" podID="9d044ebe-1e9f-4412-a94f-702501190349" containerID="c984a6a29a187f2bd2e3700a3572db64cb3aa0b921cf09c427888a1e39d63a8b" exitCode=0 Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.784284 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d057-account-create-update-kdmmk" event={"ID":"9d044ebe-1e9f-4412-a94f-702501190349","Type":"ContainerDied","Data":"c984a6a29a187f2bd2e3700a3572db64cb3aa0b921cf09c427888a1e39d63a8b"} Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.792342 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c640-account-create-update-pb7l7" event={"ID":"f8373b63-0203-47e7-b23c-714afe735109","Type":"ContainerStarted","Data":"db0e92c9f5a41f6029013e1bdb3c8547ec232497e34d12d03d0a488144286229"} Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.792392 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c640-account-create-update-pb7l7" event={"ID":"f8373b63-0203-47e7-b23c-714afe735109","Type":"ContainerStarted","Data":"eb971138e77b2668d5d58d3740d5b9f3a3f96f6b0eef6d6975d9760f2d730724"} Dec 05 09:27:58 crc kubenswrapper[4815]: I1205 09:27:58.850128 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-c640-account-create-update-pb7l7" podStartSLOduration=2.850108642 podStartE2EDuration="2.850108642s" podCreationTimestamp="2025-12-05 09:27:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:27:58.842975825 +0000 UTC m=+1277.721582662" watchObservedRunningTime="2025-12-05 09:27:58.850108642 +0000 UTC m=+1277.728715479" Dec 05 09:27:59 crc kubenswrapper[4815]: I1205 09:27:59.808244 4815 generic.go:334] "Generic (PLEG): container finished" podID="f8373b63-0203-47e7-b23c-714afe735109" containerID="db0e92c9f5a41f6029013e1bdb3c8547ec232497e34d12d03d0a488144286229" exitCode=0 Dec 05 09:27:59 crc kubenswrapper[4815]: I1205 09:27:59.808882 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c640-account-create-update-pb7l7" event={"ID":"f8373b63-0203-47e7-b23c-714afe735109","Type":"ContainerDied","Data":"db0e92c9f5a41f6029013e1bdb3c8547ec232497e34d12d03d0a488144286229"} Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.447449 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-63a3-account-create-update-5hprm" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.602759 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nkvqv\" (UniqueName: \"kubernetes.io/projected/6dfe6e19-f7c0-4360-945c-e70be5e60932-kube-api-access-nkvqv\") pod \"6dfe6e19-f7c0-4360-945c-e70be5e60932\" (UID: \"6dfe6e19-f7c0-4360-945c-e70be5e60932\") " Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.602831 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dfe6e19-f7c0-4360-945c-e70be5e60932-operator-scripts\") pod \"6dfe6e19-f7c0-4360-945c-e70be5e60932\" (UID: \"6dfe6e19-f7c0-4360-945c-e70be5e60932\") " Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.604942 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6dfe6e19-f7c0-4360-945c-e70be5e60932-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6dfe6e19-f7c0-4360-945c-e70be5e60932" (UID: "6dfe6e19-f7c0-4360-945c-e70be5e60932"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.614913 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dfe6e19-f7c0-4360-945c-e70be5e60932-kube-api-access-nkvqv" (OuterVolumeSpecName: "kube-api-access-nkvqv") pod "6dfe6e19-f7c0-4360-945c-e70be5e60932" (UID: "6dfe6e19-f7c0-4360-945c-e70be5e60932"). InnerVolumeSpecName "kube-api-access-nkvqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.656332 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2qpg9" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.711919 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nkvqv\" (UniqueName: \"kubernetes.io/projected/6dfe6e19-f7c0-4360-945c-e70be5e60932-kube-api-access-nkvqv\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.712006 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6dfe6e19-f7c0-4360-945c-e70be5e60932-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.816784 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a6fff39-63be-4b79-a9a0-dbcf8162a231-operator-scripts\") pod \"2a6fff39-63be-4b79-a9a0-dbcf8162a231\" (UID: \"2a6fff39-63be-4b79-a9a0-dbcf8162a231\") " Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.816992 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cn7fg\" (UniqueName: \"kubernetes.io/projected/2a6fff39-63be-4b79-a9a0-dbcf8162a231-kube-api-access-cn7fg\") pod \"2a6fff39-63be-4b79-a9a0-dbcf8162a231\" (UID: \"2a6fff39-63be-4b79-a9a0-dbcf8162a231\") " Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.825401 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6pbnl" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.826427 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a6fff39-63be-4b79-a9a0-dbcf8162a231-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2a6fff39-63be-4b79-a9a0-dbcf8162a231" (UID: "2a6fff39-63be-4b79-a9a0-dbcf8162a231"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.866214 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a6fff39-63be-4b79-a9a0-dbcf8162a231-kube-api-access-cn7fg" (OuterVolumeSpecName: "kube-api-access-cn7fg") pod "2a6fff39-63be-4b79-a9a0-dbcf8162a231" (UID: "2a6fff39-63be-4b79-a9a0-dbcf8162a231"). InnerVolumeSpecName "kube-api-access-cn7fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.883540 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z6t49" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.897816 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-63a3-account-create-update-5hprm" event={"ID":"6dfe6e19-f7c0-4360-945c-e70be5e60932","Type":"ContainerDied","Data":"5c28b7227af542c58bec1f7fc1608cd3cf22ddc730b8a6b3ecc9a1eeed7497ee"} Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.897858 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c28b7227af542c58bec1f7fc1608cd3cf22ddc730b8a6b3ecc9a1eeed7497ee" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.898014 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-63a3-account-create-update-5hprm" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.912989 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-2qpg9" event={"ID":"2a6fff39-63be-4b79-a9a0-dbcf8162a231","Type":"ContainerDied","Data":"396344cdb6bd79e66d62ec40c5512e37a48dce43809a9933acf89473387d2ed0"} Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.913074 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="396344cdb6bd79e66d62ec40c5512e37a48dce43809a9933acf89473387d2ed0" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.913232 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-2qpg9" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.925987 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7a77236-882a-4a90-99bc-b74679d44e01-operator-scripts\") pod \"b7a77236-882a-4a90-99bc-b74679d44e01\" (UID: \"b7a77236-882a-4a90-99bc-b74679d44e01\") " Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.926071 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vnvs\" (UniqueName: \"kubernetes.io/projected/b7a77236-882a-4a90-99bc-b74679d44e01-kube-api-access-6vnvs\") pod \"b7a77236-882a-4a90-99bc-b74679d44e01\" (UID: \"b7a77236-882a-4a90-99bc-b74679d44e01\") " Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.926135 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45462c49-fb2a-48e8-bd00-98efa3a73da7-operator-scripts\") pod \"45462c49-fb2a-48e8-bd00-98efa3a73da7\" (UID: \"45462c49-fb2a-48e8-bd00-98efa3a73da7\") " Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.926219 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk2v5\" (UniqueName: \"kubernetes.io/projected/45462c49-fb2a-48e8-bd00-98efa3a73da7-kube-api-access-fk2v5\") pod \"45462c49-fb2a-48e8-bd00-98efa3a73da7\" (UID: \"45462c49-fb2a-48e8-bd00-98efa3a73da7\") " Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.926894 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cn7fg\" (UniqueName: \"kubernetes.io/projected/2a6fff39-63be-4b79-a9a0-dbcf8162a231-kube-api-access-cn7fg\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.926925 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2a6fff39-63be-4b79-a9a0-dbcf8162a231-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.938759 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b7a77236-882a-4a90-99bc-b74679d44e01-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b7a77236-882a-4a90-99bc-b74679d44e01" (UID: "b7a77236-882a-4a90-99bc-b74679d44e01"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.938793 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/45462c49-fb2a-48e8-bd00-98efa3a73da7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "45462c49-fb2a-48e8-bd00-98efa3a73da7" (UID: "45462c49-fb2a-48e8-bd00-98efa3a73da7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.960883 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-6pbnl" Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.961333 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-6pbnl" event={"ID":"b7a77236-882a-4a90-99bc-b74679d44e01","Type":"ContainerDied","Data":"557ce497566d91e66527a2f6e3a825f6b29de51b692d2f5549058011f73a83db"} Dec 05 09:28:00 crc kubenswrapper[4815]: I1205 09:28:00.961375 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="557ce497566d91e66527a2f6e3a825f6b29de51b692d2f5549058011f73a83db" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.027563 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b7a77236-882a-4a90-99bc-b74679d44e01-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.027586 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/45462c49-fb2a-48e8-bd00-98efa3a73da7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.052826 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45462c49-fb2a-48e8-bd00-98efa3a73da7-kube-api-access-fk2v5" (OuterVolumeSpecName: "kube-api-access-fk2v5") pod "45462c49-fb2a-48e8-bd00-98efa3a73da7" (UID: "45462c49-fb2a-48e8-bd00-98efa3a73da7"). InnerVolumeSpecName "kube-api-access-fk2v5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.052930 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7a77236-882a-4a90-99bc-b74679d44e01-kube-api-access-6vnvs" (OuterVolumeSpecName: "kube-api-access-6vnvs") pod "b7a77236-882a-4a90-99bc-b74679d44e01" (UID: "b7a77236-882a-4a90-99bc-b74679d44e01"). InnerVolumeSpecName "kube-api-access-6vnvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.074756 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d057-account-create-update-kdmmk" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.128390 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d044ebe-1e9f-4412-a94f-702501190349-operator-scripts\") pod \"9d044ebe-1e9f-4412-a94f-702501190349\" (UID: \"9d044ebe-1e9f-4412-a94f-702501190349\") " Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.128446 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99gjr\" (UniqueName: \"kubernetes.io/projected/9d044ebe-1e9f-4412-a94f-702501190349-kube-api-access-99gjr\") pod \"9d044ebe-1e9f-4412-a94f-702501190349\" (UID: \"9d044ebe-1e9f-4412-a94f-702501190349\") " Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.128732 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk2v5\" (UniqueName: \"kubernetes.io/projected/45462c49-fb2a-48e8-bd00-98efa3a73da7-kube-api-access-fk2v5\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.128746 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vnvs\" (UniqueName: \"kubernetes.io/projected/b7a77236-882a-4a90-99bc-b74679d44e01-kube-api-access-6vnvs\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.129450 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d044ebe-1e9f-4412-a94f-702501190349-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9d044ebe-1e9f-4412-a94f-702501190349" (UID: "9d044ebe-1e9f-4412-a94f-702501190349"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.139239 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d044ebe-1e9f-4412-a94f-702501190349-kube-api-access-99gjr" (OuterVolumeSpecName: "kube-api-access-99gjr") pod "9d044ebe-1e9f-4412-a94f-702501190349" (UID: "9d044ebe-1e9f-4412-a94f-702501190349"). InnerVolumeSpecName "kube-api-access-99gjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.229998 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d044ebe-1e9f-4412-a94f-702501190349-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.230037 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-99gjr\" (UniqueName: \"kubernetes.io/projected/9d044ebe-1e9f-4412-a94f-702501190349-kube-api-access-99gjr\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.338646 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c640-account-create-update-pb7l7" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.533704 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wfqpb\" (UniqueName: \"kubernetes.io/projected/f8373b63-0203-47e7-b23c-714afe735109-kube-api-access-wfqpb\") pod \"f8373b63-0203-47e7-b23c-714afe735109\" (UID: \"f8373b63-0203-47e7-b23c-714afe735109\") " Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.534990 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8373b63-0203-47e7-b23c-714afe735109-operator-scripts\") pod \"f8373b63-0203-47e7-b23c-714afe735109\" (UID: \"f8373b63-0203-47e7-b23c-714afe735109\") " Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.535734 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8373b63-0203-47e7-b23c-714afe735109-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f8373b63-0203-47e7-b23c-714afe735109" (UID: "f8373b63-0203-47e7-b23c-714afe735109"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.536707 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f8373b63-0203-47e7-b23c-714afe735109-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.568123 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8373b63-0203-47e7-b23c-714afe735109-kube-api-access-wfqpb" (OuterVolumeSpecName: "kube-api-access-wfqpb") pod "f8373b63-0203-47e7-b23c-714afe735109" (UID: "f8373b63-0203-47e7-b23c-714afe735109"). InnerVolumeSpecName "kube-api-access-wfqpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.638120 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wfqpb\" (UniqueName: \"kubernetes.io/projected/f8373b63-0203-47e7-b23c-714afe735109-kube-api-access-wfqpb\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.980468 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-z6t49" event={"ID":"45462c49-fb2a-48e8-bd00-98efa3a73da7","Type":"ContainerDied","Data":"819bda1181cad3f6c7fe4c40d6c0f825ae4014f793ecd43ad41fcaa9662b16c2"} Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.980518 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="819bda1181cad3f6c7fe4c40d6c0f825ae4014f793ecd43ad41fcaa9662b16c2" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.980552 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-z6t49" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.985359 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d057-account-create-update-kdmmk" event={"ID":"9d044ebe-1e9f-4412-a94f-702501190349","Type":"ContainerDied","Data":"d1778a63e1862a7bf4b89d4601d517ec43342b0fd02c0f3817f663089dae7f3a"} Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.985388 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d1778a63e1862a7bf4b89d4601d517ec43342b0fd02c0f3817f663089dae7f3a" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.985432 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d057-account-create-update-kdmmk" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.990665 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-c640-account-create-update-pb7l7" event={"ID":"f8373b63-0203-47e7-b23c-714afe735109","Type":"ContainerDied","Data":"eb971138e77b2668d5d58d3740d5b9f3a3f96f6b0eef6d6975d9760f2d730724"} Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.990700 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb971138e77b2668d5d58d3740d5b9f3a3f96f6b0eef6d6975d9760f2d730724" Dec 05 09:28:01 crc kubenswrapper[4815]: I1205 09:28:01.990748 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-c640-account-create-update-pb7l7" Dec 05 09:28:06 crc kubenswrapper[4815]: I1205 09:28:06.034397 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bmkjh" event={"ID":"a86ad8ca-5199-41c4-ab4e-22826ea4b65d","Type":"ContainerStarted","Data":"cf891be6ee412a5b3b684ebcf7a903fdcbf3673fb15d96e38af7b05544c02c7b"} Dec 05 09:28:06 crc kubenswrapper[4815]: I1205 09:28:06.039445 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wjc6t" event={"ID":"d00433e1-638a-4b77-bd8b-9052d9c9bc11","Type":"ContainerStarted","Data":"22c4d51a44d918028a38229aaac9675684db6f88b5b01bfc62ce0fc9b2becfed"} Dec 05 09:28:06 crc kubenswrapper[4815]: I1205 09:28:06.065416 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-bmkjh" podStartSLOduration=3.150283461 podStartE2EDuration="10.065390832s" podCreationTimestamp="2025-12-05 09:27:56 +0000 UTC" firstStartedPulling="2025-12-05 09:27:58.223764662 +0000 UTC m=+1277.102371499" lastFinishedPulling="2025-12-05 09:28:05.138872033 +0000 UTC m=+1284.017478870" observedRunningTime="2025-12-05 09:28:06.059174432 +0000 UTC m=+1284.937781289" watchObservedRunningTime="2025-12-05 09:28:06.065390832 +0000 UTC m=+1284.943997669" Dec 05 09:28:17 crc kubenswrapper[4815]: I1205 09:28:17.140118 4815 generic.go:334] "Generic (PLEG): container finished" podID="a86ad8ca-5199-41c4-ab4e-22826ea4b65d" containerID="cf891be6ee412a5b3b684ebcf7a903fdcbf3673fb15d96e38af7b05544c02c7b" exitCode=0 Dec 05 09:28:17 crc kubenswrapper[4815]: I1205 09:28:17.140196 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bmkjh" event={"ID":"a86ad8ca-5199-41c4-ab4e-22826ea4b65d","Type":"ContainerDied","Data":"cf891be6ee412a5b3b684ebcf7a903fdcbf3673fb15d96e38af7b05544c02c7b"} Dec 05 09:28:17 crc kubenswrapper[4815]: I1205 09:28:17.163805 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-wjc6t" podStartSLOduration=13.113390426 podStartE2EDuration="48.163789464s" podCreationTimestamp="2025-12-05 09:27:29 +0000 UTC" firstStartedPulling="2025-12-05 09:27:30.075031576 +0000 UTC m=+1248.953638413" lastFinishedPulling="2025-12-05 09:28:05.125430614 +0000 UTC m=+1284.004037451" observedRunningTime="2025-12-05 09:28:06.081981292 +0000 UTC m=+1284.960588129" watchObservedRunningTime="2025-12-05 09:28:17.163789464 +0000 UTC m=+1296.042396301" Dec 05 09:28:18 crc kubenswrapper[4815]: I1205 09:28:18.434541 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:28:18 crc kubenswrapper[4815]: I1205 09:28:18.533118 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-combined-ca-bundle\") pod \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " Dec 05 09:28:18 crc kubenswrapper[4815]: I1205 09:28:18.533248 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntcd2\" (UniqueName: \"kubernetes.io/projected/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-kube-api-access-ntcd2\") pod \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " Dec 05 09:28:18 crc kubenswrapper[4815]: I1205 09:28:18.533280 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-config-data\") pod \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\" (UID: \"a86ad8ca-5199-41c4-ab4e-22826ea4b65d\") " Dec 05 09:28:18 crc kubenswrapper[4815]: I1205 09:28:18.540717 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-kube-api-access-ntcd2" (OuterVolumeSpecName: "kube-api-access-ntcd2") pod "a86ad8ca-5199-41c4-ab4e-22826ea4b65d" (UID: "a86ad8ca-5199-41c4-ab4e-22826ea4b65d"). InnerVolumeSpecName "kube-api-access-ntcd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:18 crc kubenswrapper[4815]: I1205 09:28:18.559032 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a86ad8ca-5199-41c4-ab4e-22826ea4b65d" (UID: "a86ad8ca-5199-41c4-ab4e-22826ea4b65d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:18 crc kubenswrapper[4815]: I1205 09:28:18.578178 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-config-data" (OuterVolumeSpecName: "config-data") pod "a86ad8ca-5199-41c4-ab4e-22826ea4b65d" (UID: "a86ad8ca-5199-41c4-ab4e-22826ea4b65d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:18 crc kubenswrapper[4815]: I1205 09:28:18.635273 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntcd2\" (UniqueName: \"kubernetes.io/projected/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-kube-api-access-ntcd2\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:18 crc kubenswrapper[4815]: I1205 09:28:18.635308 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:18 crc kubenswrapper[4815]: I1205 09:28:18.635324 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a86ad8ca-5199-41c4-ab4e-22826ea4b65d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.161856 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-bmkjh" event={"ID":"a86ad8ca-5199-41c4-ab4e-22826ea4b65d","Type":"ContainerDied","Data":"e7720da1d6348fe94047b9c8161280b90ecb01be3e8b7c29ed6e01dd0cea7293"} Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.161893 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-bmkjh" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.161896 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7720da1d6348fe94047b9c8161280b90ecb01be3e8b7c29ed6e01dd0cea7293" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.496553 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-75wjz"] Dec 05 09:28:19 crc kubenswrapper[4815]: E1205 09:28:19.497010 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dfe6e19-f7c0-4360-945c-e70be5e60932" containerName="mariadb-account-create-update" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497026 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dfe6e19-f7c0-4360-945c-e70be5e60932" containerName="mariadb-account-create-update" Dec 05 09:28:19 crc kubenswrapper[4815]: E1205 09:28:19.497051 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d044ebe-1e9f-4412-a94f-702501190349" containerName="mariadb-account-create-update" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497061 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d044ebe-1e9f-4412-a94f-702501190349" containerName="mariadb-account-create-update" Dec 05 09:28:19 crc kubenswrapper[4815]: E1205 09:28:19.497076 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45462c49-fb2a-48e8-bd00-98efa3a73da7" containerName="mariadb-database-create" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497084 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="45462c49-fb2a-48e8-bd00-98efa3a73da7" containerName="mariadb-database-create" Dec 05 09:28:19 crc kubenswrapper[4815]: E1205 09:28:19.497099 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a6fff39-63be-4b79-a9a0-dbcf8162a231" containerName="mariadb-database-create" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497106 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a6fff39-63be-4b79-a9a0-dbcf8162a231" containerName="mariadb-database-create" Dec 05 09:28:19 crc kubenswrapper[4815]: E1205 09:28:19.497122 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a86ad8ca-5199-41c4-ab4e-22826ea4b65d" containerName="keystone-db-sync" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497129 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a86ad8ca-5199-41c4-ab4e-22826ea4b65d" containerName="keystone-db-sync" Dec 05 09:28:19 crc kubenswrapper[4815]: E1205 09:28:19.497139 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7a77236-882a-4a90-99bc-b74679d44e01" containerName="mariadb-database-create" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497148 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7a77236-882a-4a90-99bc-b74679d44e01" containerName="mariadb-database-create" Dec 05 09:28:19 crc kubenswrapper[4815]: E1205 09:28:19.497169 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8373b63-0203-47e7-b23c-714afe735109" containerName="mariadb-account-create-update" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497178 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8373b63-0203-47e7-b23c-714afe735109" containerName="mariadb-account-create-update" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497435 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8373b63-0203-47e7-b23c-714afe735109" containerName="mariadb-account-create-update" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497460 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a86ad8ca-5199-41c4-ab4e-22826ea4b65d" containerName="keystone-db-sync" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497474 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7a77236-882a-4a90-99bc-b74679d44e01" containerName="mariadb-database-create" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497499 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a6fff39-63be-4b79-a9a0-dbcf8162a231" containerName="mariadb-database-create" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497515 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="45462c49-fb2a-48e8-bd00-98efa3a73da7" containerName="mariadb-database-create" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497528 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d044ebe-1e9f-4412-a94f-702501190349" containerName="mariadb-account-create-update" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.497540 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dfe6e19-f7c0-4360-945c-e70be5e60932" containerName="mariadb-account-create-update" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.498212 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.505378 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.505737 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.506316 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.506511 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.506611 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xn54c" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.506678 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-nh2hf"] Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.508479 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.530438 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-75wjz"] Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.547252 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-nh2hf"] Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.685569 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56hvd\" (UniqueName: \"kubernetes.io/projected/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-kube-api-access-56hvd\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.685866 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-nb\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.685981 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-dns-svc\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.686071 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fggsd\" (UniqueName: \"kubernetes.io/projected/d6f17f7c-378d-41a8-a230-53d3833e7d2a-kube-api-access-fggsd\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.686175 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-scripts\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.687314 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-sb\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.687654 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-credential-keys\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.687740 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-config-data\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.687818 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-combined-ca-bundle\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.688055 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-config\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.688112 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-fernet-keys\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.789333 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-scripts\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.789383 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-sb\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.789419 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-credential-keys\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.789435 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-config-data\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.789456 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-combined-ca-bundle\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.789536 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-config\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.789568 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-fernet-keys\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.789593 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56hvd\" (UniqueName: \"kubernetes.io/projected/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-kube-api-access-56hvd\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.789615 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-nb\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.789633 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-dns-svc\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.789660 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fggsd\" (UniqueName: \"kubernetes.io/projected/d6f17f7c-378d-41a8-a230-53d3833e7d2a-kube-api-access-fggsd\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.791185 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-sb\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.791629 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-config\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.791818 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-nb\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.792510 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-dns-svc\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.801994 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-4sfv2"] Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.804122 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.804624 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-credential-keys\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.805275 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-config-data\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.806281 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-brz2d" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.806899 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.810793 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-fernet-keys\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.812254 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-combined-ca-bundle\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.812482 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-scripts\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.824352 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-4sfv2"] Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.830674 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fggsd\" (UniqueName: \"kubernetes.io/projected/d6f17f7c-378d-41a8-a230-53d3833e7d2a-kube-api-access-fggsd\") pod \"dnsmasq-dns-8fb4c6755-nh2hf\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.832297 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56hvd\" (UniqueName: \"kubernetes.io/projected/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-kube-api-access-56hvd\") pod \"keystone-bootstrap-75wjz\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.840362 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.897662 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltz2l\" (UniqueName: \"kubernetes.io/projected/664c6884-a071-4e07-a90a-14146b1bb46e-kube-api-access-ltz2l\") pod \"barbican-db-sync-4sfv2\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.897815 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-combined-ca-bundle\") pod \"barbican-db-sync-4sfv2\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.897855 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-db-sync-config-data\") pod \"barbican-db-sync-4sfv2\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.933248 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-96bs8"] Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.934753 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.961001 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.961226 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.961415 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5z58g" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.989317 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-dgjpz"] Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.995598 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.999056 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-cwk9h" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.999264 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 09:28:19 crc kubenswrapper[4815]: I1205 09:28:19.999396 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.000419 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-combined-ca-bundle\") pod \"barbican-db-sync-4sfv2\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.000464 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7vzd\" (UniqueName: \"kubernetes.io/projected/c832b179-b3a7-45d6-a360-11e05100d1f2-kube-api-access-m7vzd\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.000509 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-db-sync-config-data\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.000560 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-db-sync-config-data\") pod \"barbican-db-sync-4sfv2\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.000587 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-scripts\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.000617 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c832b179-b3a7-45d6-a360-11e05100d1f2-etc-machine-id\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.000644 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-config-data\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.000682 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltz2l\" (UniqueName: \"kubernetes.io/projected/664c6884-a071-4e07-a90a-14146b1bb46e-kube-api-access-ltz2l\") pod \"barbican-db-sync-4sfv2\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.000742 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-combined-ca-bundle\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.010460 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-combined-ca-bundle\") pod \"barbican-db-sync-4sfv2\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.018447 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-96bs8"] Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.021353 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-db-sync-config-data\") pod \"barbican-db-sync-4sfv2\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.053560 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-dgjpz"] Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.066263 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltz2l\" (UniqueName: \"kubernetes.io/projected/664c6884-a071-4e07-a90a-14146b1bb46e-kube-api-access-ltz2l\") pod \"barbican-db-sync-4sfv2\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.105565 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7vzd\" (UniqueName: \"kubernetes.io/projected/c832b179-b3a7-45d6-a360-11e05100d1f2-kube-api-access-m7vzd\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.105610 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-db-sync-config-data\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.105650 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npqd8\" (UniqueName: \"kubernetes.io/projected/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-kube-api-access-npqd8\") pod \"neutron-db-sync-dgjpz\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.106482 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-scripts\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.106527 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c832b179-b3a7-45d6-a360-11e05100d1f2-etc-machine-id\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.106563 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-config-data\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.106598 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c832b179-b3a7-45d6-a360-11e05100d1f2-etc-machine-id\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.106661 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-config\") pod \"neutron-db-sync-dgjpz\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.106689 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-combined-ca-bundle\") pod \"neutron-db-sync-dgjpz\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.106724 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-combined-ca-bundle\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.114724 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-combined-ca-bundle\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.114912 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-scripts\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.115663 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-config-data\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.118078 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-db-sync-config-data\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.120195 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.131390 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-ss9wk"] Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.142131 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.163851 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ss9wk"] Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.208618 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-config-data\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.208901 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-combined-ca-bundle\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.208981 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npqd8\" (UniqueName: \"kubernetes.io/projected/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-kube-api-access-npqd8\") pod \"neutron-db-sync-dgjpz\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.209054 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-scripts\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.209076 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c66190b5-87b0-440b-9894-1bffb2cfca3f-logs\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.209103 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cftk\" (UniqueName: \"kubernetes.io/projected/c66190b5-87b0-440b-9894-1bffb2cfca3f-kube-api-access-4cftk\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.209132 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-config\") pod \"neutron-db-sync-dgjpz\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.209157 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-combined-ca-bundle\") pod \"neutron-db-sync-dgjpz\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.216039 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-combined-ca-bundle\") pod \"neutron-db-sync-dgjpz\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.221249 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.221422 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-dgw9g" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.224688 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-config\") pod \"neutron-db-sync-dgjpz\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.255090 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.283233 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7vzd\" (UniqueName: \"kubernetes.io/projected/c832b179-b3a7-45d6-a360-11e05100d1f2-kube-api-access-m7vzd\") pod \"cinder-db-sync-96bs8\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.301956 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.311879 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-config-data\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.311941 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-combined-ca-bundle\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.312024 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-scripts\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.312048 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c66190b5-87b0-440b-9894-1bffb2cfca3f-logs\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.312073 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cftk\" (UniqueName: \"kubernetes.io/projected/c66190b5-87b0-440b-9894-1bffb2cfca3f-kube-api-access-4cftk\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.315912 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c66190b5-87b0-440b-9894-1bffb2cfca3f-logs\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.327528 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-scripts\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.330192 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-combined-ca-bundle\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.333545 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-config-data\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.355099 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npqd8\" (UniqueName: \"kubernetes.io/projected/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-kube-api-access-npqd8\") pod \"neutron-db-sync-dgjpz\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.372322 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cftk\" (UniqueName: \"kubernetes.io/projected/c66190b5-87b0-440b-9894-1bffb2cfca3f-kube-api-access-4cftk\") pod \"placement-db-sync-ss9wk\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.402943 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-96bs8" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.415664 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-nh2hf"] Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.422689 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.442581 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.451659 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.467028 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.467327 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.479648 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b559c875f-bl7vh"] Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.481004 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.504858 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.522979 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kgmv\" (UniqueName: \"kubernetes.io/projected/501b099d-0ffb-4c12-9809-15107cd1431a-kube-api-access-5kgmv\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.523028 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-config-data\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.523148 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-dns-svc\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.523181 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.523202 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-config\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.523218 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-run-httpd\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.523233 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.523262 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.523297 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.523320 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-log-httpd\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.523341 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-scripts\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.523358 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5thc\" (UniqueName: \"kubernetes.io/projected/786b442d-b215-4707-b9b5-62f167b4299d-kube-api-access-b5thc\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.526617 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b559c875f-bl7vh"] Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.540406 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ss9wk" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626449 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626505 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-config\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626524 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-run-httpd\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626540 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626620 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626660 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626683 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-log-httpd\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626712 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-scripts\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626731 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5thc\" (UniqueName: \"kubernetes.io/projected/786b442d-b215-4707-b9b5-62f167b4299d-kube-api-access-b5thc\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626757 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kgmv\" (UniqueName: \"kubernetes.io/projected/501b099d-0ffb-4c12-9809-15107cd1431a-kube-api-access-5kgmv\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626789 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-config-data\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.626839 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-dns-svc\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.633453 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-nb\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.633570 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-config\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.633783 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-run-httpd\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.635008 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-sb\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.636900 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-dns-svc\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.652968 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-log-httpd\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.668413 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-scripts\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.670893 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-config-data\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.700886 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kgmv\" (UniqueName: \"kubernetes.io/projected/501b099d-0ffb-4c12-9809-15107cd1431a-kube-api-access-5kgmv\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.701373 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.715518 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5thc\" (UniqueName: \"kubernetes.io/projected/786b442d-b215-4707-b9b5-62f167b4299d-kube-api-access-b5thc\") pod \"dnsmasq-dns-6b559c875f-bl7vh\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.716244 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.783908 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.814887 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:20 crc kubenswrapper[4815]: I1205 09:28:20.934918 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-nh2hf"] Dec 05 09:28:21 crc kubenswrapper[4815]: I1205 09:28:21.260032 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" event={"ID":"d6f17f7c-378d-41a8-a230-53d3833e7d2a","Type":"ContainerStarted","Data":"610f9f21907e3b58123ecbd280d7b775ae814e27dcc9059fcdd04feab426cc5c"} Dec 05 09:28:21 crc kubenswrapper[4815]: I1205 09:28:21.523997 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-ss9wk"] Dec 05 09:28:21 crc kubenswrapper[4815]: W1205 09:28:21.529719 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc66190b5_87b0_440b_9894_1bffb2cfca3f.slice/crio-b19745a03816abf498bc255d946b1d525a681a6224e1fec8dbccceb3214716b1 WatchSource:0}: Error finding container b19745a03816abf498bc255d946b1d525a681a6224e1fec8dbccceb3214716b1: Status 404 returned error can't find the container with id b19745a03816abf498bc255d946b1d525a681a6224e1fec8dbccceb3214716b1 Dec 05 09:28:21 crc kubenswrapper[4815]: I1205 09:28:21.543252 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-96bs8"] Dec 05 09:28:21 crc kubenswrapper[4815]: I1205 09:28:21.591935 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-75wjz"] Dec 05 09:28:21 crc kubenswrapper[4815]: I1205 09:28:21.733657 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-4sfv2"] Dec 05 09:28:21 crc kubenswrapper[4815]: I1205 09:28:21.758152 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-dgjpz"] Dec 05 09:28:21 crc kubenswrapper[4815]: I1205 09:28:21.912805 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b559c875f-bl7vh"] Dec 05 09:28:21 crc kubenswrapper[4815]: I1205 09:28:21.947861 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:28:22 crc kubenswrapper[4815]: I1205 09:28:22.270864 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501b099d-0ffb-4c12-9809-15107cd1431a","Type":"ContainerStarted","Data":"3eecbebffc2f5140c795577bcc53a3f686652af35b0a1f83ffeb99fce0a3a55c"} Dec 05 09:28:22 crc kubenswrapper[4815]: I1205 09:28:22.272677 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-75wjz" event={"ID":"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b","Type":"ContainerStarted","Data":"1f7ef1fbe6aa6478447e7bacb1027bef23a8b2e202b8cd1622fc1db108b8ce74"} Dec 05 09:28:22 crc kubenswrapper[4815]: I1205 09:28:22.272705 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-75wjz" event={"ID":"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b","Type":"ContainerStarted","Data":"cd3312f0860fd7569b14ee2ded65b75a441b6ce80972e4c88804a817a3e307e7"} Dec 05 09:28:22 crc kubenswrapper[4815]: I1205 09:28:22.275352 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-4sfv2" event={"ID":"664c6884-a071-4e07-a90a-14146b1bb46e","Type":"ContainerStarted","Data":"a384372f62dc06996e95a203ce0e42cd7cacb30019f7d1519e59e7c3638dacf4"} Dec 05 09:28:22 crc kubenswrapper[4815]: I1205 09:28:22.276949 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dgjpz" event={"ID":"a2ea4b20-281a-43e1-b6af-c2c209b10e9d","Type":"ContainerStarted","Data":"3bcb739c62ee00862983222f7bb32b4fc2a7ea16f98dd9b0159baa4726e19f8d"} Dec 05 09:28:22 crc kubenswrapper[4815]: I1205 09:28:22.277879 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ss9wk" event={"ID":"c66190b5-87b0-440b-9894-1bffb2cfca3f","Type":"ContainerStarted","Data":"b19745a03816abf498bc255d946b1d525a681a6224e1fec8dbccceb3214716b1"} Dec 05 09:28:22 crc kubenswrapper[4815]: I1205 09:28:22.278899 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" event={"ID":"786b442d-b215-4707-b9b5-62f167b4299d","Type":"ContainerStarted","Data":"125d5bebe4c4680dc609e350bb9bc16f73deb17b00c83c90033ed16b693d0831"} Dec 05 09:28:22 crc kubenswrapper[4815]: I1205 09:28:22.286878 4815 generic.go:334] "Generic (PLEG): container finished" podID="d6f17f7c-378d-41a8-a230-53d3833e7d2a" containerID="838728ed25765c17f6d4a00f9f1909534a6fe10f9267e4492e4e9a95a04d7929" exitCode=0 Dec 05 09:28:22 crc kubenswrapper[4815]: I1205 09:28:22.287084 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" event={"ID":"d6f17f7c-378d-41a8-a230-53d3833e7d2a","Type":"ContainerDied","Data":"838728ed25765c17f6d4a00f9f1909534a6fe10f9267e4492e4e9a95a04d7929"} Dec 05 09:28:22 crc kubenswrapper[4815]: I1205 09:28:22.290169 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-96bs8" event={"ID":"c832b179-b3a7-45d6-a360-11e05100d1f2","Type":"ContainerStarted","Data":"47b3b5497a244b810b78019cb0f3cbb0180745c82ef73c332c8adbfd2ff7de13"} Dec 05 09:28:22 crc kubenswrapper[4815]: I1205 09:28:22.294549 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-75wjz" podStartSLOduration=3.294533889 podStartE2EDuration="3.294533889s" podCreationTimestamp="2025-12-05 09:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:28:22.294196479 +0000 UTC m=+1301.172803316" watchObservedRunningTime="2025-12-05 09:28:22.294533889 +0000 UTC m=+1301.173140726" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.664595 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.720830 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-sb\") pod \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.721896 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-config\") pod \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.721938 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-nb\") pod \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.721982 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-dns-svc\") pod \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.722061 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fggsd\" (UniqueName: \"kubernetes.io/projected/d6f17f7c-378d-41a8-a230-53d3833e7d2a-kube-api-access-fggsd\") pod \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\" (UID: \"d6f17f7c-378d-41a8-a230-53d3833e7d2a\") " Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.761922 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6f17f7c-378d-41a8-a230-53d3833e7d2a-kube-api-access-fggsd" (OuterVolumeSpecName: "kube-api-access-fggsd") pod "d6f17f7c-378d-41a8-a230-53d3833e7d2a" (UID: "d6f17f7c-378d-41a8-a230-53d3833e7d2a"). InnerVolumeSpecName "kube-api-access-fggsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.823756 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fggsd\" (UniqueName: \"kubernetes.io/projected/d6f17f7c-378d-41a8-a230-53d3833e7d2a-kube-api-access-fggsd\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.846237 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d6f17f7c-378d-41a8-a230-53d3833e7d2a" (UID: "d6f17f7c-378d-41a8-a230-53d3833e7d2a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.846587 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-config" (OuterVolumeSpecName: "config") pod "d6f17f7c-378d-41a8-a230-53d3833e7d2a" (UID: "d6f17f7c-378d-41a8-a230-53d3833e7d2a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.861412 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d6f17f7c-378d-41a8-a230-53d3833e7d2a" (UID: "d6f17f7c-378d-41a8-a230-53d3833e7d2a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.925558 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.925589 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.925599 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.930829 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d6f17f7c-378d-41a8-a230-53d3833e7d2a" (UID: "d6f17f7c-378d-41a8-a230-53d3833e7d2a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:22.935900 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:23.027365 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6f17f7c-378d-41a8-a230-53d3833e7d2a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:23.305251 4815 generic.go:334] "Generic (PLEG): container finished" podID="786b442d-b215-4707-b9b5-62f167b4299d" containerID="d40f4852bd467be3c730273e865e2158952ef153741e7255587aea05fc2cad13" exitCode=0 Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:23.305479 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" event={"ID":"786b442d-b215-4707-b9b5-62f167b4299d","Type":"ContainerDied","Data":"d40f4852bd467be3c730273e865e2158952ef153741e7255587aea05fc2cad13"} Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:23.324801 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" event={"ID":"d6f17f7c-378d-41a8-a230-53d3833e7d2a","Type":"ContainerDied","Data":"610f9f21907e3b58123ecbd280d7b775ae814e27dcc9059fcdd04feab426cc5c"} Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:23.324864 4815 scope.go:117] "RemoveContainer" containerID="838728ed25765c17f6d4a00f9f1909534a6fe10f9267e4492e4e9a95a04d7929" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:23.325001 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8fb4c6755-nh2hf" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:23.368639 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dgjpz" event={"ID":"a2ea4b20-281a-43e1-b6af-c2c209b10e9d","Type":"ContainerStarted","Data":"d09086a87a983214a6a3881c0aa175532e712f484214c0057dc7f8fea016f095"} Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:23.447981 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-dgjpz" podStartSLOduration=4.447959362 podStartE2EDuration="4.447959362s" podCreationTimestamp="2025-12-05 09:28:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:28:23.430472126 +0000 UTC m=+1302.309078963" watchObservedRunningTime="2025-12-05 09:28:23.447959362 +0000 UTC m=+1302.326566199" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:23.521427 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-nh2hf"] Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:23.578809 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8fb4c6755-nh2hf"] Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:25.427842 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6f17f7c-378d-41a8-a230-53d3833e7d2a" path="/var/lib/kubelet/pods/d6f17f7c-378d-41a8-a230-53d3833e7d2a/volumes" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:27.405627 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" event={"ID":"786b442d-b215-4707-b9b5-62f167b4299d","Type":"ContainerStarted","Data":"da191c1e5444641bf2024a83cbb365f9b4665a5f8054d499b49d976a2cea2baf"} Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:27.406019 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:30 crc kubenswrapper[4815]: I1205 09:28:27.432976 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" podStartSLOduration=7.432954827 podStartE2EDuration="7.432954827s" podCreationTimestamp="2025-12-05 09:28:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:28:27.426127889 +0000 UTC m=+1306.304734726" watchObservedRunningTime="2025-12-05 09:28:27.432954827 +0000 UTC m=+1306.311561664" Dec 05 09:28:35 crc kubenswrapper[4815]: I1205 09:28:35.817960 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:28:35 crc kubenswrapper[4815]: I1205 09:28:35.897310 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-mvfhk"] Dec 05 09:28:35 crc kubenswrapper[4815]: I1205 09:28:35.897579 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" podUID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" containerName="dnsmasq-dns" containerID="cri-o://6e63a5434a480287d59e55a265a98a2afe7af7dc9da69f0494958e9ecc809fce" gracePeriod=10 Dec 05 09:28:37 crc kubenswrapper[4815]: I1205 09:28:37.400264 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" podUID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Dec 05 09:28:41 crc kubenswrapper[4815]: I1205 09:28:41.531052 4815 generic.go:334] "Generic (PLEG): container finished" podID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" containerID="6e63a5434a480287d59e55a265a98a2afe7af7dc9da69f0494958e9ecc809fce" exitCode=0 Dec 05 09:28:41 crc kubenswrapper[4815]: I1205 09:28:41.531121 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" event={"ID":"f38535a2-7fda-4ebc-9d88-241703ec1a1b","Type":"ContainerDied","Data":"6e63a5434a480287d59e55a265a98a2afe7af7dc9da69f0494958e9ecc809fce"} Dec 05 09:28:42 crc kubenswrapper[4815]: I1205 09:28:42.400174 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" podUID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Dec 05 09:28:45 crc kubenswrapper[4815]: I1205 09:28:45.570734 4815 generic.go:334] "Generic (PLEG): container finished" podID="ae3babbb-a8ff-4653-bd0f-0e414ea1f95b" containerID="1f7ef1fbe6aa6478447e7bacb1027bef23a8b2e202b8cd1622fc1db108b8ce74" exitCode=0 Dec 05 09:28:45 crc kubenswrapper[4815]: I1205 09:28:45.570841 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-75wjz" event={"ID":"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b","Type":"ContainerDied","Data":"1f7ef1fbe6aa6478447e7bacb1027bef23a8b2e202b8cd1622fc1db108b8ce74"} Dec 05 09:28:45 crc kubenswrapper[4815]: E1205 09:28:45.627394 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2" Dec 05 09:28:45 crc kubenswrapper[4815]: E1205 09:28:45.627646 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m7vzd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-96bs8_openstack(c832b179-b3a7-45d6-a360-11e05100d1f2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:28:45 crc kubenswrapper[4815]: E1205 09:28:45.629114 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-96bs8" podUID="c832b179-b3a7-45d6-a360-11e05100d1f2" Dec 05 09:28:46 crc kubenswrapper[4815]: E1205 09:28:46.581300 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:7a2056615520e272bae43ec3f34e2ba7a92c1d364b8d9106b53bd694619fc9c2\\\"\"" pod="openstack/cinder-db-sync-96bs8" podUID="c832b179-b3a7-45d6-a360-11e05100d1f2" Dec 05 09:28:47 crc kubenswrapper[4815]: I1205 09:28:47.400400 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" podUID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.109:5353: connect: connection refused" Dec 05 09:28:47 crc kubenswrapper[4815]: I1205 09:28:47.400562 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:28:48 crc kubenswrapper[4815]: I1205 09:28:48.601128 4815 generic.go:334] "Generic (PLEG): container finished" podID="d00433e1-638a-4b77-bd8b-9052d9c9bc11" containerID="22c4d51a44d918028a38229aaac9675684db6f88b5b01bfc62ce0fc9b2becfed" exitCode=0 Dec 05 09:28:48 crc kubenswrapper[4815]: I1205 09:28:48.601197 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wjc6t" event={"ID":"d00433e1-638a-4b77-bd8b-9052d9c9bc11","Type":"ContainerDied","Data":"22c4d51a44d918028a38229aaac9675684db6f88b5b01bfc62ce0fc9b2becfed"} Dec 05 09:28:49 crc kubenswrapper[4815]: E1205 09:28:49.071570 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api@sha256:f24234939afca841e46ea4d17bec959b63705ab0e75476465e777d44905c5f1b" Dec 05 09:28:49 crc kubenswrapper[4815]: E1205 09:28:49.071792 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api@sha256:f24234939afca841e46ea4d17bec959b63705ab0e75476465e777d44905c5f1b,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4cftk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-ss9wk_openstack(c66190b5-87b0-440b-9894-1bffb2cfca3f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:28:49 crc kubenswrapper[4815]: E1205 09:28:49.074668 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-ss9wk" podUID="c66190b5-87b0-440b-9894-1bffb2cfca3f" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.139467 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.313155 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-scripts\") pod \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.313237 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-combined-ca-bundle\") pod \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.313296 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-fernet-keys\") pod \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.313333 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-config-data\") pod \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.313438 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56hvd\" (UniqueName: \"kubernetes.io/projected/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-kube-api-access-56hvd\") pod \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.313552 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-credential-keys\") pod \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\" (UID: \"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b\") " Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.320003 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-scripts" (OuterVolumeSpecName: "scripts") pod "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b" (UID: "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.320846 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-kube-api-access-56hvd" (OuterVolumeSpecName: "kube-api-access-56hvd") pod "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b" (UID: "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b"). InnerVolumeSpecName "kube-api-access-56hvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.323302 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b" (UID: "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.328690 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b" (UID: "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.349173 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b" (UID: "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.362638 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-config-data" (OuterVolumeSpecName: "config-data") pod "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b" (UID: "ae3babbb-a8ff-4653-bd0f-0e414ea1f95b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.418528 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56hvd\" (UniqueName: \"kubernetes.io/projected/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-kube-api-access-56hvd\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.418568 4815 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.418582 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.418594 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.418604 4815 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.418615 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:49 crc kubenswrapper[4815]: E1205 09:28:49.531032 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:43a24796dabde68270dbfefa107205e173fdd6a0dc701502858cadbede69da31" Dec 05 09:28:49 crc kubenswrapper[4815]: E1205 09:28:49.531199 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central@sha256:43a24796dabde68270dbfefa107205e173fdd6a0dc701502858cadbede69da31,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d6h56fh545h67fh5fdh577h86h65bh68ch8dh59dhdbh76h664hb4hd9h586h54hdbh685h75h695h5ddh5ch5cdh64h597h85hdfh5cdh57ch589q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5kgmv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(501b099d-0ffb-4c12-9809-15107cd1431a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.533906 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.580031 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.623168 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46zjb\" (UniqueName: \"kubernetes.io/projected/f38535a2-7fda-4ebc-9d88-241703ec1a1b-kube-api-access-46zjb\") pod \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.623288 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-nb\") pod \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.623337 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-config\") pod \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.623363 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-dns-svc\") pod \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.623461 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-sb\") pod \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\" (UID: \"f38535a2-7fda-4ebc-9d88-241703ec1a1b\") " Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.628824 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f38535a2-7fda-4ebc-9d88-241703ec1a1b-kube-api-access-46zjb" (OuterVolumeSpecName: "kube-api-access-46zjb") pod "f38535a2-7fda-4ebc-9d88-241703ec1a1b" (UID: "f38535a2-7fda-4ebc-9d88-241703ec1a1b"). InnerVolumeSpecName "kube-api-access-46zjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.640365 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" event={"ID":"f38535a2-7fda-4ebc-9d88-241703ec1a1b","Type":"ContainerDied","Data":"dcdb4a28c86cc7090d555b60bf024153c553f0988c1cd40ad3235f0e2ad80f51"} Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.640456 4815 scope.go:117] "RemoveContainer" containerID="6e63a5434a480287d59e55a265a98a2afe7af7dc9da69f0494958e9ecc809fce" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.640644 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-db7757ddc-mvfhk" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.670272 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-75wjz" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.671047 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-75wjz" event={"ID":"ae3babbb-a8ff-4653-bd0f-0e414ea1f95b","Type":"ContainerDied","Data":"cd3312f0860fd7569b14ee2ded65b75a441b6ce80972e4c88804a817a3e307e7"} Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.671083 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cd3312f0860fd7569b14ee2ded65b75a441b6ce80972e4c88804a817a3e307e7" Dec 05 09:28:49 crc kubenswrapper[4815]: E1205 09:28:49.677260 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api@sha256:f24234939afca841e46ea4d17bec959b63705ab0e75476465e777d44905c5f1b\\\"\"" pod="openstack/placement-db-sync-ss9wk" podUID="c66190b5-87b0-440b-9894-1bffb2cfca3f" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.699120 4815 scope.go:117] "RemoveContainer" containerID="8fd74d008a79136a2ac470f0d1d17844e69664d5d1c34c645fd8b9fd0c1e51b0" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.722712 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f38535a2-7fda-4ebc-9d88-241703ec1a1b" (UID: "f38535a2-7fda-4ebc-9d88-241703ec1a1b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.725242 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46zjb\" (UniqueName: \"kubernetes.io/projected/f38535a2-7fda-4ebc-9d88-241703ec1a1b-kube-api-access-46zjb\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.725274 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.726953 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f38535a2-7fda-4ebc-9d88-241703ec1a1b" (UID: "f38535a2-7fda-4ebc-9d88-241703ec1a1b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.731940 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-config" (OuterVolumeSpecName: "config") pod "f38535a2-7fda-4ebc-9d88-241703ec1a1b" (UID: "f38535a2-7fda-4ebc-9d88-241703ec1a1b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.748855 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f38535a2-7fda-4ebc-9d88-241703ec1a1b" (UID: "f38535a2-7fda-4ebc-9d88-241703ec1a1b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.825944 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.825980 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.825989 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f38535a2-7fda-4ebc-9d88-241703ec1a1b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.979861 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-mvfhk"] Dec 05 09:28:49 crc kubenswrapper[4815]: I1205 09:28:49.988145 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-db7757ddc-mvfhk"] Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.002097 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wjc6t" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.031790 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-combined-ca-bundle\") pod \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.034204 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-db-sync-config-data\") pod \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.034459 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-config-data\") pod \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.034623 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k46vx\" (UniqueName: \"kubernetes.io/projected/d00433e1-638a-4b77-bd8b-9052d9c9bc11-kube-api-access-k46vx\") pod \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\" (UID: \"d00433e1-638a-4b77-bd8b-9052d9c9bc11\") " Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.042522 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d00433e1-638a-4b77-bd8b-9052d9c9bc11-kube-api-access-k46vx" (OuterVolumeSpecName: "kube-api-access-k46vx") pod "d00433e1-638a-4b77-bd8b-9052d9c9bc11" (UID: "d00433e1-638a-4b77-bd8b-9052d9c9bc11"). InnerVolumeSpecName "kube-api-access-k46vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.048076 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d00433e1-638a-4b77-bd8b-9052d9c9bc11" (UID: "d00433e1-638a-4b77-bd8b-9052d9c9bc11"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.065447 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d00433e1-638a-4b77-bd8b-9052d9c9bc11" (UID: "d00433e1-638a-4b77-bd8b-9052d9c9bc11"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.115136 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-config-data" (OuterVolumeSpecName: "config-data") pod "d00433e1-638a-4b77-bd8b-9052d9c9bc11" (UID: "d00433e1-638a-4b77-bd8b-9052d9c9bc11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.136222 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.136255 4815 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.136264 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d00433e1-638a-4b77-bd8b-9052d9c9bc11-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.136275 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k46vx\" (UniqueName: \"kubernetes.io/projected/d00433e1-638a-4b77-bd8b-9052d9c9bc11-kube-api-access-k46vx\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.230948 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-75wjz"] Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.238077 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-75wjz"] Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.333600 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-59q99"] Dec 05 09:28:50 crc kubenswrapper[4815]: E1205 09:28:50.334041 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d00433e1-638a-4b77-bd8b-9052d9c9bc11" containerName="glance-db-sync" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.334068 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d00433e1-638a-4b77-bd8b-9052d9c9bc11" containerName="glance-db-sync" Dec 05 09:28:50 crc kubenswrapper[4815]: E1205 09:28:50.334093 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae3babbb-a8ff-4653-bd0f-0e414ea1f95b" containerName="keystone-bootstrap" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.334099 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae3babbb-a8ff-4653-bd0f-0e414ea1f95b" containerName="keystone-bootstrap" Dec 05 09:28:50 crc kubenswrapper[4815]: E1205 09:28:50.334113 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" containerName="dnsmasq-dns" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.334119 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" containerName="dnsmasq-dns" Dec 05 09:28:50 crc kubenswrapper[4815]: E1205 09:28:50.334128 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f17f7c-378d-41a8-a230-53d3833e7d2a" containerName="init" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.334134 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f17f7c-378d-41a8-a230-53d3833e7d2a" containerName="init" Dec 05 09:28:50 crc kubenswrapper[4815]: E1205 09:28:50.334145 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" containerName="init" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.334150 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" containerName="init" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.334352 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae3babbb-a8ff-4653-bd0f-0e414ea1f95b" containerName="keystone-bootstrap" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.334370 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d00433e1-638a-4b77-bd8b-9052d9c9bc11" containerName="glance-db-sync" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.334380 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" containerName="dnsmasq-dns" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.334387 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6f17f7c-378d-41a8-a230-53d3833e7d2a" containerName="init" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.336275 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.338540 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.338965 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.339104 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.339214 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.341212 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-combined-ca-bundle\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.341296 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-486bl\" (UniqueName: \"kubernetes.io/projected/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-kube-api-access-486bl\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.341318 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-scripts\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.341376 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-credential-keys\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.341436 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-config-data\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.341547 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-fernet-keys\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.341586 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-59q99"] Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.345520 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xn54c" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.442910 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-fernet-keys\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.442993 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-combined-ca-bundle\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.443026 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-486bl\" (UniqueName: \"kubernetes.io/projected/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-kube-api-access-486bl\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.443044 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-scripts\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.443110 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-credential-keys\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.443192 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-config-data\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.447943 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-credential-keys\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.448337 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-combined-ca-bundle\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.448819 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-fernet-keys\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.451960 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-scripts\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.457290 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-config-data\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.472637 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-486bl\" (UniqueName: \"kubernetes.io/projected/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-kube-api-access-486bl\") pod \"keystone-bootstrap-59q99\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.659404 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.682114 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-4sfv2" event={"ID":"664c6884-a071-4e07-a90a-14146b1bb46e","Type":"ContainerStarted","Data":"0bd36f2c0a25614d8e3b243d5764d3f2b92f0ea7796d9d90abf1961264d67e81"} Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.683946 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wjc6t" event={"ID":"d00433e1-638a-4b77-bd8b-9052d9c9bc11","Type":"ContainerDied","Data":"a8caca550a24792c339fb0c3c512a9e1842dbacda8abc64154575e41e25678b7"} Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.683997 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8caca550a24792c339fb0c3c512a9e1842dbacda8abc64154575e41e25678b7" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.684075 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wjc6t" Dec 05 09:28:50 crc kubenswrapper[4815]: I1205 09:28:50.699855 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-4sfv2" podStartSLOduration=3.973907646 podStartE2EDuration="31.699832389s" podCreationTimestamp="2025-12-05 09:28:19 +0000 UTC" firstStartedPulling="2025-12-05 09:28:21.798058322 +0000 UTC m=+1300.676665159" lastFinishedPulling="2025-12-05 09:28:49.523983065 +0000 UTC m=+1328.402589902" observedRunningTime="2025-12-05 09:28:50.697028067 +0000 UTC m=+1329.575634924" watchObservedRunningTime="2025-12-05 09:28:50.699832389 +0000 UTC m=+1329.578439226" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.155414 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-g2rdw"] Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.157631 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.169954 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rf5z6\" (UniqueName: \"kubernetes.io/projected/aa2e6ba1-185a-465b-9174-f2e8ee425cca-kube-api-access-rf5z6\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.170040 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-sb\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.170063 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-dns-svc\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.170134 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-nb\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.170158 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-config\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.275571 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-sb\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.275622 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-dns-svc\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.275703 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-nb\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.275730 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-config\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.275807 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rf5z6\" (UniqueName: \"kubernetes.io/projected/aa2e6ba1-185a-465b-9174-f2e8ee425cca-kube-api-access-rf5z6\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.289581 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-g2rdw"] Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.290666 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-nb\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.290731 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-dns-svc\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.291700 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-config\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.294210 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-sb\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.334202 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rf5z6\" (UniqueName: \"kubernetes.io/projected/aa2e6ba1-185a-465b-9174-f2e8ee425cca-kube-api-access-rf5z6\") pod \"dnsmasq-dns-66cd6b8b67-g2rdw\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.440897 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae3babbb-a8ff-4653-bd0f-0e414ea1f95b" path="/var/lib/kubelet/pods/ae3babbb-a8ff-4653-bd0f-0e414ea1f95b/volumes" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.441728 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f38535a2-7fda-4ebc-9d88-241703ec1a1b" path="/var/lib/kubelet/pods/f38535a2-7fda-4ebc-9d88-241703ec1a1b/volumes" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.603216 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.608827 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-59q99"] Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.708955 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501b099d-0ffb-4c12-9809-15107cd1431a","Type":"ContainerStarted","Data":"7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792"} Dec 05 09:28:51 crc kubenswrapper[4815]: I1205 09:28:51.717766 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-59q99" event={"ID":"1b7b7063-a8d5-41bd-b969-72035e0f2c6b","Type":"ContainerStarted","Data":"486a89bb5fb5c6557600e338d15bf90228622758cfd900531f403a3237661cd0"} Dec 05 09:28:52 crc kubenswrapper[4815]: I1205 09:28:52.136854 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-g2rdw"] Dec 05 09:28:52 crc kubenswrapper[4815]: W1205 09:28:52.147274 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa2e6ba1_185a_465b_9174_f2e8ee425cca.slice/crio-caa8effca42a6a423740e96f4351b58bdb6d485adb08d224384adee4626ad41b WatchSource:0}: Error finding container caa8effca42a6a423740e96f4351b58bdb6d485adb08d224384adee4626ad41b: Status 404 returned error can't find the container with id caa8effca42a6a423740e96f4351b58bdb6d485adb08d224384adee4626ad41b Dec 05 09:28:52 crc kubenswrapper[4815]: I1205 09:28:52.727882 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-59q99" event={"ID":"1b7b7063-a8d5-41bd-b969-72035e0f2c6b","Type":"ContainerStarted","Data":"d460b7b15ce45f9ce88ec5a4df9c5bfd875de498a3ecb6315d71b8d6fc916d84"} Dec 05 09:28:52 crc kubenswrapper[4815]: I1205 09:28:52.732207 4815 generic.go:334] "Generic (PLEG): container finished" podID="aa2e6ba1-185a-465b-9174-f2e8ee425cca" containerID="e4e6b4e27dbdb7355e89be90c17df7777bf5b65d3b4e5457125f12032cb349cd" exitCode=0 Dec 05 09:28:52 crc kubenswrapper[4815]: I1205 09:28:52.732265 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" event={"ID":"aa2e6ba1-185a-465b-9174-f2e8ee425cca","Type":"ContainerDied","Data":"e4e6b4e27dbdb7355e89be90c17df7777bf5b65d3b4e5457125f12032cb349cd"} Dec 05 09:28:52 crc kubenswrapper[4815]: I1205 09:28:52.732296 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" event={"ID":"aa2e6ba1-185a-465b-9174-f2e8ee425cca","Type":"ContainerStarted","Data":"caa8effca42a6a423740e96f4351b58bdb6d485adb08d224384adee4626ad41b"} Dec 05 09:28:52 crc kubenswrapper[4815]: I1205 09:28:52.790596 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-59q99" podStartSLOduration=2.7905738319999998 podStartE2EDuration="2.790573832s" podCreationTimestamp="2025-12-05 09:28:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:28:52.752037206 +0000 UTC m=+1331.630644043" watchObservedRunningTime="2025-12-05 09:28:52.790573832 +0000 UTC m=+1331.669180669" Dec 05 09:28:53 crc kubenswrapper[4815]: I1205 09:28:53.748120 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" event={"ID":"aa2e6ba1-185a-465b-9174-f2e8ee425cca","Type":"ContainerStarted","Data":"6849430cddef0d2968a8412f853dbb9cfb66fbedf6166d0d163d2938749deb2e"} Dec 05 09:28:53 crc kubenswrapper[4815]: I1205 09:28:53.748610 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:28:53 crc kubenswrapper[4815]: I1205 09:28:53.779609 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" podStartSLOduration=2.779591241 podStartE2EDuration="2.779591241s" podCreationTimestamp="2025-12-05 09:28:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:28:53.777183212 +0000 UTC m=+1332.655790059" watchObservedRunningTime="2025-12-05 09:28:53.779591241 +0000 UTC m=+1332.658198078" Dec 05 09:28:55 crc kubenswrapper[4815]: I1205 09:28:55.773301 4815 generic.go:334] "Generic (PLEG): container finished" podID="664c6884-a071-4e07-a90a-14146b1bb46e" containerID="0bd36f2c0a25614d8e3b243d5764d3f2b92f0ea7796d9d90abf1961264d67e81" exitCode=0 Dec 05 09:28:55 crc kubenswrapper[4815]: I1205 09:28:55.773392 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-4sfv2" event={"ID":"664c6884-a071-4e07-a90a-14146b1bb46e","Type":"ContainerDied","Data":"0bd36f2c0a25614d8e3b243d5764d3f2b92f0ea7796d9d90abf1961264d67e81"} Dec 05 09:28:57 crc kubenswrapper[4815]: I1205 09:28:57.791314 4815 generic.go:334] "Generic (PLEG): container finished" podID="1b7b7063-a8d5-41bd-b969-72035e0f2c6b" containerID="d460b7b15ce45f9ce88ec5a4df9c5bfd875de498a3ecb6315d71b8d6fc916d84" exitCode=0 Dec 05 09:28:57 crc kubenswrapper[4815]: I1205 09:28:57.791428 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-59q99" event={"ID":"1b7b7063-a8d5-41bd-b969-72035e0f2c6b","Type":"ContainerDied","Data":"d460b7b15ce45f9ce88ec5a4df9c5bfd875de498a3ecb6315d71b8d6fc916d84"} Dec 05 09:28:58 crc kubenswrapper[4815]: I1205 09:28:58.805458 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-4sfv2" event={"ID":"664c6884-a071-4e07-a90a-14146b1bb46e","Type":"ContainerDied","Data":"a384372f62dc06996e95a203ce0e42cd7cacb30019f7d1519e59e7c3638dacf4"} Dec 05 09:28:58 crc kubenswrapper[4815]: I1205 09:28:58.805780 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a384372f62dc06996e95a203ce0e42cd7cacb30019f7d1519e59e7c3638dacf4" Dec 05 09:28:58 crc kubenswrapper[4815]: I1205 09:28:58.860010 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.034464 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-db-sync-config-data\") pod \"664c6884-a071-4e07-a90a-14146b1bb46e\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.034594 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltz2l\" (UniqueName: \"kubernetes.io/projected/664c6884-a071-4e07-a90a-14146b1bb46e-kube-api-access-ltz2l\") pod \"664c6884-a071-4e07-a90a-14146b1bb46e\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.034737 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-combined-ca-bundle\") pod \"664c6884-a071-4e07-a90a-14146b1bb46e\" (UID: \"664c6884-a071-4e07-a90a-14146b1bb46e\") " Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.045122 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/664c6884-a071-4e07-a90a-14146b1bb46e-kube-api-access-ltz2l" (OuterVolumeSpecName: "kube-api-access-ltz2l") pod "664c6884-a071-4e07-a90a-14146b1bb46e" (UID: "664c6884-a071-4e07-a90a-14146b1bb46e"). InnerVolumeSpecName "kube-api-access-ltz2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.076840 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "664c6884-a071-4e07-a90a-14146b1bb46e" (UID: "664c6884-a071-4e07-a90a-14146b1bb46e"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.080421 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "664c6884-a071-4e07-a90a-14146b1bb46e" (UID: "664c6884-a071-4e07-a90a-14146b1bb46e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.137382 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.137443 4815 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/664c6884-a071-4e07-a90a-14146b1bb46e-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.137458 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltz2l\" (UniqueName: \"kubernetes.io/projected/664c6884-a071-4e07-a90a-14146b1bb46e-kube-api-access-ltz2l\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.222009 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.237776 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-credential-keys\") pod \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.237829 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-486bl\" (UniqueName: \"kubernetes.io/projected/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-kube-api-access-486bl\") pod \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.237861 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-scripts\") pod \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.237898 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-config-data\") pod \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.237921 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-combined-ca-bundle\") pod \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.237956 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-fernet-keys\") pod \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\" (UID: \"1b7b7063-a8d5-41bd-b969-72035e0f2c6b\") " Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.242626 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-scripts" (OuterVolumeSpecName: "scripts") pod "1b7b7063-a8d5-41bd-b969-72035e0f2c6b" (UID: "1b7b7063-a8d5-41bd-b969-72035e0f2c6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.247379 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "1b7b7063-a8d5-41bd-b969-72035e0f2c6b" (UID: "1b7b7063-a8d5-41bd-b969-72035e0f2c6b"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.273944 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "1b7b7063-a8d5-41bd-b969-72035e0f2c6b" (UID: "1b7b7063-a8d5-41bd-b969-72035e0f2c6b"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.275662 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-kube-api-access-486bl" (OuterVolumeSpecName: "kube-api-access-486bl") pod "1b7b7063-a8d5-41bd-b969-72035e0f2c6b" (UID: "1b7b7063-a8d5-41bd-b969-72035e0f2c6b"). InnerVolumeSpecName "kube-api-access-486bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.305713 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b7b7063-a8d5-41bd-b969-72035e0f2c6b" (UID: "1b7b7063-a8d5-41bd-b969-72035e0f2c6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.321075 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-config-data" (OuterVolumeSpecName: "config-data") pod "1b7b7063-a8d5-41bd-b969-72035e0f2c6b" (UID: "1b7b7063-a8d5-41bd-b969-72035e0f2c6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.342033 4815 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.342069 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-486bl\" (UniqueName: \"kubernetes.io/projected/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-kube-api-access-486bl\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.342084 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.342094 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.342104 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.342113 4815 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/1b7b7063-a8d5-41bd-b969-72035e0f2c6b-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.815403 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-96bs8" event={"ID":"c832b179-b3a7-45d6-a360-11e05100d1f2","Type":"ContainerStarted","Data":"c86c9131a647365e140b7d211616e5b78f71e312cf85293d078ba9093f0e52da"} Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.817745 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501b099d-0ffb-4c12-9809-15107cd1431a","Type":"ContainerStarted","Data":"19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d"} Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.819285 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-59q99" event={"ID":"1b7b7063-a8d5-41bd-b969-72035e0f2c6b","Type":"ContainerDied","Data":"486a89bb5fb5c6557600e338d15bf90228622758cfd900531f403a3237661cd0"} Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.819313 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-4sfv2" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.819325 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="486a89bb5fb5c6557600e338d15bf90228622758cfd900531f403a3237661cd0" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.819296 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-59q99" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.854830 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-96bs8" podStartSLOduration=3.2793801399999998 podStartE2EDuration="40.854808555s" podCreationTimestamp="2025-12-05 09:28:19 +0000 UTC" firstStartedPulling="2025-12-05 09:28:21.567895413 +0000 UTC m=+1300.446502250" lastFinishedPulling="2025-12-05 09:28:59.143323828 +0000 UTC m=+1338.021930665" observedRunningTime="2025-12-05 09:28:59.852432796 +0000 UTC m=+1338.731039633" watchObservedRunningTime="2025-12-05 09:28:59.854808555 +0000 UTC m=+1338.733415392" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.943470 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-6f56547bd4-2dxhd"] Dec 05 09:28:59 crc kubenswrapper[4815]: E1205 09:28:59.943825 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b7b7063-a8d5-41bd-b969-72035e0f2c6b" containerName="keystone-bootstrap" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.943846 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b7b7063-a8d5-41bd-b969-72035e0f2c6b" containerName="keystone-bootstrap" Dec 05 09:28:59 crc kubenswrapper[4815]: E1205 09:28:59.943912 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="664c6884-a071-4e07-a90a-14146b1bb46e" containerName="barbican-db-sync" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.943919 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="664c6884-a071-4e07-a90a-14146b1bb46e" containerName="barbican-db-sync" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.944082 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b7b7063-a8d5-41bd-b969-72035e0f2c6b" containerName="keystone-bootstrap" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.944101 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="664c6884-a071-4e07-a90a-14146b1bb46e" containerName="barbican-db-sync" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.944672 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.946884 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.946889 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.947310 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-xn54c" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.947433 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.947517 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 05 09:28:59 crc kubenswrapper[4815]: I1205 09:28:59.949195 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.008769 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6f56547bd4-2dxhd"] Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.051962 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfzrm\" (UniqueName: \"kubernetes.io/projected/7518e451-ec15-4df9-8fa0-842ca6094b36-kube-api-access-zfzrm\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.052056 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-public-tls-certs\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.052106 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-credential-keys\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.052141 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-combined-ca-bundle\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.052208 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-fernet-keys\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.052237 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-scripts\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.052349 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-config-data\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.052390 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-internal-tls-certs\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.153650 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-public-tls-certs\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.153721 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-credential-keys\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.153762 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-combined-ca-bundle\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.153802 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-fernet-keys\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.153830 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-scripts\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.153873 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-config-data\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.153896 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-internal-tls-certs\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.153975 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfzrm\" (UniqueName: \"kubernetes.io/projected/7518e451-ec15-4df9-8fa0-842ca6094b36-kube-api-access-zfzrm\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.161574 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-scripts\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.161998 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-combined-ca-bundle\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.162568 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-fernet-keys\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.162804 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-config-data\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.164088 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-credential-keys\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.171049 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-public-tls-certs\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.185730 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7518e451-ec15-4df9-8fa0-842ca6094b36-internal-tls-certs\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.188723 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfzrm\" (UniqueName: \"kubernetes.io/projected/7518e451-ec15-4df9-8fa0-842ca6094b36-kube-api-access-zfzrm\") pod \"keystone-6f56547bd4-2dxhd\" (UID: \"7518e451-ec15-4df9-8fa0-842ca6094b36\") " pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.262353 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.286669 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-57db4d6df7-gdwss"] Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.288701 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.298301 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.298624 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-brz2d" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.298784 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.312884 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-57db4d6df7-gdwss"] Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.358617 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-596686cb9d-rwjpw"] Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.360181 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.370800 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.413628 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-596686cb9d-rwjpw"] Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.464368 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92f1b7e3-1999-448d-90c1-9845b162701b-config-data-custom\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.464411 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9lvrk\" (UniqueName: \"kubernetes.io/projected/92f1b7e3-1999-448d-90c1-9845b162701b-kube-api-access-9lvrk\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.464463 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f638151-76e9-429e-b283-d741ad857276-logs\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.464506 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f1b7e3-1999-448d-90c1-9845b162701b-combined-ca-bundle\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.464533 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f1b7e3-1999-448d-90c1-9845b162701b-config-data\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.464671 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f638151-76e9-429e-b283-d741ad857276-config-data\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.464708 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f638151-76e9-429e-b283-d741ad857276-config-data-custom\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.464780 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92f1b7e3-1999-448d-90c1-9845b162701b-logs\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.464890 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f638151-76e9-429e-b283-d741ad857276-combined-ca-bundle\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.464916 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh2gl\" (UniqueName: \"kubernetes.io/projected/3f638151-76e9-429e-b283-d741ad857276-kube-api-access-rh2gl\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.491257 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-g2rdw"] Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.491477 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" podUID="aa2e6ba1-185a-465b-9174-f2e8ee425cca" containerName="dnsmasq-dns" containerID="cri-o://6849430cddef0d2968a8412f853dbb9cfb66fbedf6166d0d163d2938749deb2e" gracePeriod=10 Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.495873 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.573831 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92f1b7e3-1999-448d-90c1-9845b162701b-logs\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.574160 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f638151-76e9-429e-b283-d741ad857276-combined-ca-bundle\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.574179 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh2gl\" (UniqueName: \"kubernetes.io/projected/3f638151-76e9-429e-b283-d741ad857276-kube-api-access-rh2gl\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.574206 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92f1b7e3-1999-448d-90c1-9845b162701b-config-data-custom\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.574223 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9lvrk\" (UniqueName: \"kubernetes.io/projected/92f1b7e3-1999-448d-90c1-9845b162701b-kube-api-access-9lvrk\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.574250 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f638151-76e9-429e-b283-d741ad857276-logs\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.574267 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f1b7e3-1999-448d-90c1-9845b162701b-combined-ca-bundle\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.574320 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f1b7e3-1999-448d-90c1-9845b162701b-config-data\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.574355 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f638151-76e9-429e-b283-d741ad857276-config-data\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.574383 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f638151-76e9-429e-b283-d741ad857276-config-data-custom\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.575405 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85487d4b67-wqr5k"] Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.577196 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.580025 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92f1b7e3-1999-448d-90c1-9845b162701b-config-data-custom\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.582899 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3f638151-76e9-429e-b283-d741ad857276-config-data-custom\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.583146 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3f638151-76e9-429e-b283-d741ad857276-logs\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.583343 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92f1b7e3-1999-448d-90c1-9845b162701b-logs\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.587170 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f1b7e3-1999-448d-90c1-9845b162701b-combined-ca-bundle\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.588680 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85487d4b67-wqr5k"] Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.590815 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3f638151-76e9-429e-b283-d741ad857276-config-data\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.601335 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f1b7e3-1999-448d-90c1-9845b162701b-config-data\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.602964 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f638151-76e9-429e-b283-d741ad857276-combined-ca-bundle\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.606332 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9lvrk\" (UniqueName: \"kubernetes.io/projected/92f1b7e3-1999-448d-90c1-9845b162701b-kube-api-access-9lvrk\") pod \"barbican-worker-57db4d6df7-gdwss\" (UID: \"92f1b7e3-1999-448d-90c1-9845b162701b\") " pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.618713 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh2gl\" (UniqueName: \"kubernetes.io/projected/3f638151-76e9-429e-b283-d741ad857276-kube-api-access-rh2gl\") pod \"barbican-keystone-listener-596686cb9d-rwjpw\" (UID: \"3f638151-76e9-429e-b283-d741ad857276\") " pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.673272 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-cd8fb7c76-9g4zb"] Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.692030 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.696863 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.737738 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cd8fb7c76-9g4zb"] Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.740473 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-57db4d6df7-gdwss" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.773411 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.785513 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsk25\" (UniqueName: \"kubernetes.io/projected/36611e8e-67fb-420a-a1d6-1ad80e671aa9-kube-api-access-qsk25\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.785593 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-sb\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.785744 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-nb\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.785977 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-combined-ca-bundle\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.789280 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c428r\" (UniqueName: \"kubernetes.io/projected/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-kube-api-access-c428r\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.789402 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-dns-svc\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.789432 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-logs\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.789465 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data-custom\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.789623 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.789670 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-config\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.898771 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-nb\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.898930 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-combined-ca-bundle\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.898964 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c428r\" (UniqueName: \"kubernetes.io/projected/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-kube-api-access-c428r\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.899005 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-dns-svc\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.899022 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-logs\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.899046 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data-custom\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.899125 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.899154 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-config\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.899189 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsk25\" (UniqueName: \"kubernetes.io/projected/36611e8e-67fb-420a-a1d6-1ad80e671aa9-kube-api-access-qsk25\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.899685 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-logs\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.899219 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-sb\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.900808 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-nb\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.901779 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-sb\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.902284 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-dns-svc\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.903334 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-config\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.914817 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.920218 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-combined-ca-bundle\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.924440 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c428r\" (UniqueName: \"kubernetes.io/projected/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-kube-api-access-c428r\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.927736 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data-custom\") pod \"barbican-api-cd8fb7c76-9g4zb\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.930617 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsk25\" (UniqueName: \"kubernetes.io/projected/36611e8e-67fb-420a-a1d6-1ad80e671aa9-kube-api-access-qsk25\") pod \"dnsmasq-dns-85487d4b67-wqr5k\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.934010 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.934396 4815 generic.go:334] "Generic (PLEG): container finished" podID="aa2e6ba1-185a-465b-9174-f2e8ee425cca" containerID="6849430cddef0d2968a8412f853dbb9cfb66fbedf6166d0d163d2938749deb2e" exitCode=0 Dec 05 09:29:00 crc kubenswrapper[4815]: I1205 09:29:00.934454 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" event={"ID":"aa2e6ba1-185a-465b-9174-f2e8ee425cca","Type":"ContainerDied","Data":"6849430cddef0d2968a8412f853dbb9cfb66fbedf6166d0d163d2938749deb2e"} Dec 05 09:29:01 crc kubenswrapper[4815]: I1205 09:29:01.098374 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:01 crc kubenswrapper[4815]: I1205 09:29:01.304957 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-6f56547bd4-2dxhd"] Dec 05 09:29:01 crc kubenswrapper[4815]: I1205 09:29:01.385961 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-596686cb9d-rwjpw"] Dec 05 09:29:01 crc kubenswrapper[4815]: I1205 09:29:01.487914 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-57db4d6df7-gdwss"] Dec 05 09:29:01 crc kubenswrapper[4815]: I1205 09:29:01.714998 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85487d4b67-wqr5k"] Dec 05 09:29:01 crc kubenswrapper[4815]: I1205 09:29:01.865643 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:29:01 crc kubenswrapper[4815]: I1205 09:29:01.973690 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-config\") pod \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " Dec 05 09:29:01 crc kubenswrapper[4815]: I1205 09:29:01.973758 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-dns-svc\") pod \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " Dec 05 09:29:01 crc kubenswrapper[4815]: I1205 09:29:01.973883 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-nb\") pod \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " Dec 05 09:29:01 crc kubenswrapper[4815]: I1205 09:29:01.973998 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rf5z6\" (UniqueName: \"kubernetes.io/projected/aa2e6ba1-185a-465b-9174-f2e8ee425cca-kube-api-access-rf5z6\") pod \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " Dec 05 09:29:01 crc kubenswrapper[4815]: I1205 09:29:01.974076 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-sb\") pod \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\" (UID: \"aa2e6ba1-185a-465b-9174-f2e8ee425cca\") " Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.005663 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa2e6ba1-185a-465b-9174-f2e8ee425cca-kube-api-access-rf5z6" (OuterVolumeSpecName: "kube-api-access-rf5z6") pod "aa2e6ba1-185a-465b-9174-f2e8ee425cca" (UID: "aa2e6ba1-185a-465b-9174-f2e8ee425cca"). InnerVolumeSpecName "kube-api-access-rf5z6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.039012 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" event={"ID":"3f638151-76e9-429e-b283-d741ad857276","Type":"ContainerStarted","Data":"a70fffbda22432c4e724b7006e435461d95ff81065f49bbffa67064269d9b9e1"} Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.048686 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57db4d6df7-gdwss" event={"ID":"92f1b7e3-1999-448d-90c1-9845b162701b","Type":"ContainerStarted","Data":"04a55ceee197b42068c7d06462a5cfbc6be2f4abce1df4420647264459dab613"} Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.068167 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6f56547bd4-2dxhd" event={"ID":"7518e451-ec15-4df9-8fa0-842ca6094b36","Type":"ContainerStarted","Data":"e7d799a15f0f1e31b840bc7a3243c27e04624b33054da264ca9ec15c35760723"} Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.068665 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-6f56547bd4-2dxhd" event={"ID":"7518e451-ec15-4df9-8fa0-842ca6094b36","Type":"ContainerStarted","Data":"cf38c3e1672079221c2cec5c7cad864350fb7ac9c6d31dfd517141055651dc21"} Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.069982 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.080756 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rf5z6\" (UniqueName: \"kubernetes.io/projected/aa2e6ba1-185a-465b-9174-f2e8ee425cca-kube-api-access-rf5z6\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.083212 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" event={"ID":"aa2e6ba1-185a-465b-9174-f2e8ee425cca","Type":"ContainerDied","Data":"caa8effca42a6a423740e96f4351b58bdb6d485adb08d224384adee4626ad41b"} Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.083256 4815 scope.go:117] "RemoveContainer" containerID="6849430cddef0d2968a8412f853dbb9cfb66fbedf6166d0d163d2938749deb2e" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.083373 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.093660 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" event={"ID":"36611e8e-67fb-420a-a1d6-1ad80e671aa9","Type":"ContainerStarted","Data":"5e8297460545b0eeb384e86dad603fb98370ea66556995921e11f01952f98f37"} Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.113031 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-config" (OuterVolumeSpecName: "config") pod "aa2e6ba1-185a-465b-9174-f2e8ee425cca" (UID: "aa2e6ba1-185a-465b-9174-f2e8ee425cca"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.116634 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-6f56547bd4-2dxhd" podStartSLOduration=3.116613586 podStartE2EDuration="3.116613586s" podCreationTimestamp="2025-12-05 09:28:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:02.091255081 +0000 UTC m=+1340.969861918" watchObservedRunningTime="2025-12-05 09:29:02.116613586 +0000 UTC m=+1340.995220423" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.117824 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-cd8fb7c76-9g4zb"] Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.199461 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.203449 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "aa2e6ba1-185a-465b-9174-f2e8ee425cca" (UID: "aa2e6ba1-185a-465b-9174-f2e8ee425cca"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.209411 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "aa2e6ba1-185a-465b-9174-f2e8ee425cca" (UID: "aa2e6ba1-185a-465b-9174-f2e8ee425cca"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.230472 4815 scope.go:117] "RemoveContainer" containerID="e4e6b4e27dbdb7355e89be90c17df7777bf5b65d3b4e5457125f12032cb349cd" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.241624 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "aa2e6ba1-185a-465b-9174-f2e8ee425cca" (UID: "aa2e6ba1-185a-465b-9174-f2e8ee425cca"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.301154 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.301193 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.301205 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/aa2e6ba1-185a-465b-9174-f2e8ee425cca-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.443204 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-g2rdw"] Dec 05 09:29:02 crc kubenswrapper[4815]: E1205 09:29:02.444048 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod36611e8e_67fb_420a_a1d6_1ad80e671aa9.slice/crio-conmon-ee6b35a57697661062488bf6f8ea1bce91b1b01c51fa23980199e963b8b5eb7f.scope\": RecentStats: unable to find data in memory cache]" Dec 05 09:29:02 crc kubenswrapper[4815]: I1205 09:29:02.452278 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66cd6b8b67-g2rdw"] Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.123060 4815 generic.go:334] "Generic (PLEG): container finished" podID="36611e8e-67fb-420a-a1d6-1ad80e671aa9" containerID="ee6b35a57697661062488bf6f8ea1bce91b1b01c51fa23980199e963b8b5eb7f" exitCode=0 Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.123560 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" event={"ID":"36611e8e-67fb-420a-a1d6-1ad80e671aa9","Type":"ContainerDied","Data":"ee6b35a57697661062488bf6f8ea1bce91b1b01c51fa23980199e963b8b5eb7f"} Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.140608 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ss9wk" event={"ID":"c66190b5-87b0-440b-9894-1bffb2cfca3f","Type":"ContainerStarted","Data":"ac13c63149b7164d533764020d964e0736c593645c983b6e58b63ac587517a86"} Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.220052 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cd8fb7c76-9g4zb" event={"ID":"2acc91d0-1a89-4cef-bcbb-90bb61e1323f","Type":"ContainerStarted","Data":"fba393df4f5ab9c99d2166ead3e4b4b3180ea1a121161d1a36e462c3245f3a2a"} Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.220101 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cd8fb7c76-9g4zb" event={"ID":"2acc91d0-1a89-4cef-bcbb-90bb61e1323f","Type":"ContainerStarted","Data":"45ef5ea6df8164df3bad60882166ae470ca705e908389d422fa6fc2185fe4ea0"} Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.220113 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cd8fb7c76-9g4zb" event={"ID":"2acc91d0-1a89-4cef-bcbb-90bb61e1323f","Type":"ContainerStarted","Data":"9064091118df2d26038dca148e2c9cd75684bfcd298947a82a2ff47b82c049f4"} Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.220156 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.220184 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.283087 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-ss9wk" podStartSLOduration=2.775356011 podStartE2EDuration="43.283064147s" podCreationTimestamp="2025-12-05 09:28:20 +0000 UTC" firstStartedPulling="2025-12-05 09:28:21.547605804 +0000 UTC m=+1300.426212641" lastFinishedPulling="2025-12-05 09:29:02.05531394 +0000 UTC m=+1340.933920777" observedRunningTime="2025-12-05 09:29:03.216237951 +0000 UTC m=+1342.094844788" watchObservedRunningTime="2025-12-05 09:29:03.283064147 +0000 UTC m=+1342.161670984" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.302643 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podStartSLOduration=3.302623474 podStartE2EDuration="3.302623474s" podCreationTimestamp="2025-12-05 09:29:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:03.272227383 +0000 UTC m=+1342.150834220" watchObservedRunningTime="2025-12-05 09:29:03.302623474 +0000 UTC m=+1342.181230311" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.454539 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa2e6ba1-185a-465b-9174-f2e8ee425cca" path="/var/lib/kubelet/pods/aa2e6ba1-185a-465b-9174-f2e8ee425cca/volumes" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.642202 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-588897d4d8-8nmk9"] Dec 05 09:29:03 crc kubenswrapper[4815]: E1205 09:29:03.642577 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa2e6ba1-185a-465b-9174-f2e8ee425cca" containerName="init" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.642594 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa2e6ba1-185a-465b-9174-f2e8ee425cca" containerName="init" Dec 05 09:29:03 crc kubenswrapper[4815]: E1205 09:29:03.642616 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa2e6ba1-185a-465b-9174-f2e8ee425cca" containerName="dnsmasq-dns" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.642621 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa2e6ba1-185a-465b-9174-f2e8ee425cca" containerName="dnsmasq-dns" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.642821 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa2e6ba1-185a-465b-9174-f2e8ee425cca" containerName="dnsmasq-dns" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.643682 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.646840 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.657071 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-588897d4d8-8nmk9"] Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.661833 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.816634 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-combined-ca-bundle\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.816873 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7266a34-7fa1-442b-9e83-e66665a5cd12-logs\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.817059 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-config-data-custom\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.817110 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-config-data\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.817179 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-public-tls-certs\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.817363 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-internal-tls-certs\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.817474 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-975s8\" (UniqueName: \"kubernetes.io/projected/d7266a34-7fa1-442b-9e83-e66665a5cd12-kube-api-access-975s8\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.920042 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-internal-tls-certs\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.920100 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-975s8\" (UniqueName: \"kubernetes.io/projected/d7266a34-7fa1-442b-9e83-e66665a5cd12-kube-api-access-975s8\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.920162 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-combined-ca-bundle\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.920477 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7266a34-7fa1-442b-9e83-e66665a5cd12-logs\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.920591 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-config-data-custom\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.920624 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-config-data\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.920654 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-public-tls-certs\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.921358 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d7266a34-7fa1-442b-9e83-e66665a5cd12-logs\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.926839 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-config-data-custom\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.930519 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-config-data\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.933059 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-public-tls-certs\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.934393 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-combined-ca-bundle\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.938469 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d7266a34-7fa1-442b-9e83-e66665a5cd12-internal-tls-certs\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:03 crc kubenswrapper[4815]: I1205 09:29:03.985391 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-975s8\" (UniqueName: \"kubernetes.io/projected/d7266a34-7fa1-442b-9e83-e66665a5cd12-kube-api-access-975s8\") pod \"barbican-api-588897d4d8-8nmk9\" (UID: \"d7266a34-7fa1-442b-9e83-e66665a5cd12\") " pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:04 crc kubenswrapper[4815]: I1205 09:29:04.266232 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:06 crc kubenswrapper[4815]: I1205 09:29:06.037391 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-588897d4d8-8nmk9"] Dec 05 09:29:06 crc kubenswrapper[4815]: W1205 09:29:06.045879 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7266a34_7fa1_442b_9e83_e66665a5cd12.slice/crio-0d93506e257e8469e6407796ea6550922047b09cf89c4c2582e828f6d5c2f297 WatchSource:0}: Error finding container 0d93506e257e8469e6407796ea6550922047b09cf89c4c2582e828f6d5c2f297: Status 404 returned error can't find the container with id 0d93506e257e8469e6407796ea6550922047b09cf89c4c2582e828f6d5c2f297 Dec 05 09:29:06 crc kubenswrapper[4815]: I1205 09:29:06.245744 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" event={"ID":"36611e8e-67fb-420a-a1d6-1ad80e671aa9","Type":"ContainerStarted","Data":"25adc14737ef33dd8e88043f554f20a7ecad51e478b30ff09764cfc3541be9a2"} Dec 05 09:29:06 crc kubenswrapper[4815]: I1205 09:29:06.245893 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:06 crc kubenswrapper[4815]: I1205 09:29:06.247341 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" event={"ID":"3f638151-76e9-429e-b283-d741ad857276","Type":"ContainerStarted","Data":"9c40791b597bba5c5094cd05bd1395dda6584cc49b206934366c2336382cfeaf"} Dec 05 09:29:06 crc kubenswrapper[4815]: I1205 09:29:06.247368 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" event={"ID":"3f638151-76e9-429e-b283-d741ad857276","Type":"ContainerStarted","Data":"118d5cf25daea1899bccb36535131dc11d88dcbba113ba02c84b72c6cedf2e2b"} Dec 05 09:29:06 crc kubenswrapper[4815]: I1205 09:29:06.249096 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57db4d6df7-gdwss" event={"ID":"92f1b7e3-1999-448d-90c1-9845b162701b","Type":"ContainerStarted","Data":"c20184092824a2bbc7c629a2caeec355961359cdd35a47419175f202c68048ff"} Dec 05 09:29:06 crc kubenswrapper[4815]: I1205 09:29:06.249141 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-57db4d6df7-gdwss" event={"ID":"92f1b7e3-1999-448d-90c1-9845b162701b","Type":"ContainerStarted","Data":"0849a11b162eb9b68707e0c52fb15d535e25380f53bad43e295a5a125f0418f8"} Dec 05 09:29:06 crc kubenswrapper[4815]: I1205 09:29:06.250668 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-588897d4d8-8nmk9" event={"ID":"d7266a34-7fa1-442b-9e83-e66665a5cd12","Type":"ContainerStarted","Data":"0d93506e257e8469e6407796ea6550922047b09cf89c4c2582e828f6d5c2f297"} Dec 05 09:29:06 crc kubenswrapper[4815]: I1205 09:29:06.291021 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" podStartSLOduration=6.290999828 podStartE2EDuration="6.290999828s" podCreationTimestamp="2025-12-05 09:29:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:06.271905315 +0000 UTC m=+1345.150512152" watchObservedRunningTime="2025-12-05 09:29:06.290999828 +0000 UTC m=+1345.169606665" Dec 05 09:29:06 crc kubenswrapper[4815]: I1205 09:29:06.604584 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-66cd6b8b67-g2rdw" podUID="aa2e6ba1-185a-465b-9174-f2e8ee425cca" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.135:5353: i/o timeout" Dec 05 09:29:07 crc kubenswrapper[4815]: I1205 09:29:07.267687 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-588897d4d8-8nmk9" event={"ID":"d7266a34-7fa1-442b-9e83-e66665a5cd12","Type":"ContainerStarted","Data":"b7f5fecb8ae585d728cc8fba0e06ccd6d3d1c58f4dd7d428b8717c016e69024a"} Dec 05 09:29:07 crc kubenswrapper[4815]: I1205 09:29:07.285387 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-57db4d6df7-gdwss" podStartSLOduration=3.223183172 podStartE2EDuration="7.285367843s" podCreationTimestamp="2025-12-05 09:29:00 +0000 UTC" firstStartedPulling="2025-12-05 09:29:01.485940931 +0000 UTC m=+1340.364547768" lastFinishedPulling="2025-12-05 09:29:05.548125602 +0000 UTC m=+1344.426732439" observedRunningTime="2025-12-05 09:29:06.290944477 +0000 UTC m=+1345.169551314" watchObservedRunningTime="2025-12-05 09:29:07.285367843 +0000 UTC m=+1346.163974680" Dec 05 09:29:07 crc kubenswrapper[4815]: I1205 09:29:07.292551 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-596686cb9d-rwjpw" podStartSLOduration=3.226669784 podStartE2EDuration="7.292530531s" podCreationTimestamp="2025-12-05 09:29:00 +0000 UTC" firstStartedPulling="2025-12-05 09:29:01.451419391 +0000 UTC m=+1340.330026218" lastFinishedPulling="2025-12-05 09:29:05.517280128 +0000 UTC m=+1344.395886965" observedRunningTime="2025-12-05 09:29:07.282461138 +0000 UTC m=+1346.161067975" watchObservedRunningTime="2025-12-05 09:29:07.292530531 +0000 UTC m=+1346.171137378" Dec 05 09:29:10 crc kubenswrapper[4815]: I1205 09:29:10.935670 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:11 crc kubenswrapper[4815]: I1205 09:29:11.005109 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b559c875f-bl7vh"] Dec 05 09:29:11 crc kubenswrapper[4815]: I1205 09:29:11.005392 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" podUID="786b442d-b215-4707-b9b5-62f167b4299d" containerName="dnsmasq-dns" containerID="cri-o://da191c1e5444641bf2024a83cbb365f9b4665a5f8054d499b49d976a2cea2baf" gracePeriod=10 Dec 05 09:29:12 crc kubenswrapper[4815]: I1205 09:29:12.182803 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.140:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:29:12 crc kubenswrapper[4815]: I1205 09:29:12.183032 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.140:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:29:13 crc kubenswrapper[4815]: I1205 09:29:13.315864 4815 generic.go:334] "Generic (PLEG): container finished" podID="786b442d-b215-4707-b9b5-62f167b4299d" containerID="da191c1e5444641bf2024a83cbb365f9b4665a5f8054d499b49d976a2cea2baf" exitCode=0 Dec 05 09:29:13 crc kubenswrapper[4815]: I1205 09:29:13.316020 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" event={"ID":"786b442d-b215-4707-b9b5-62f167b4299d","Type":"ContainerDied","Data":"da191c1e5444641bf2024a83cbb365f9b4665a5f8054d499b49d976a2cea2baf"} Dec 05 09:29:15 crc kubenswrapper[4815]: I1205 09:29:15.816727 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" podUID="786b442d-b215-4707-b9b5-62f167b4299d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Dec 05 09:29:16 crc kubenswrapper[4815]: I1205 09:29:16.182767 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.140:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:29:16 crc kubenswrapper[4815]: I1205 09:29:16.182795 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.140:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:29:17 crc kubenswrapper[4815]: I1205 09:29:17.264729 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.140:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:29:17 crc kubenswrapper[4815]: I1205 09:29:17.264726 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.140:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:29:19 crc kubenswrapper[4815]: I1205 09:29:19.203146 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:29:19 crc kubenswrapper[4815]: I1205 09:29:19.288238 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.193276 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.193601 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.381434 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.386882 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" event={"ID":"786b442d-b215-4707-b9b5-62f167b4299d","Type":"ContainerDied","Data":"125d5bebe4c4680dc609e350bb9bc16f73deb17b00c83c90033ed16b693d0831"} Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.386924 4815 scope.go:117] "RemoveContainer" containerID="da191c1e5444641bf2024a83cbb365f9b4665a5f8054d499b49d976a2cea2baf" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.387030 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b559c875f-bl7vh" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.514321 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-sb\") pod \"786b442d-b215-4707-b9b5-62f167b4299d\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.514431 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-nb\") pod \"786b442d-b215-4707-b9b5-62f167b4299d\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.514468 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-dns-svc\") pod \"786b442d-b215-4707-b9b5-62f167b4299d\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.514592 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b5thc\" (UniqueName: \"kubernetes.io/projected/786b442d-b215-4707-b9b5-62f167b4299d-kube-api-access-b5thc\") pod \"786b442d-b215-4707-b9b5-62f167b4299d\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.514619 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-config\") pod \"786b442d-b215-4707-b9b5-62f167b4299d\" (UID: \"786b442d-b215-4707-b9b5-62f167b4299d\") " Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.524586 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/786b442d-b215-4707-b9b5-62f167b4299d-kube-api-access-b5thc" (OuterVolumeSpecName: "kube-api-access-b5thc") pod "786b442d-b215-4707-b9b5-62f167b4299d" (UID: "786b442d-b215-4707-b9b5-62f167b4299d"). InnerVolumeSpecName "kube-api-access-b5thc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.567753 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "786b442d-b215-4707-b9b5-62f167b4299d" (UID: "786b442d-b215-4707-b9b5-62f167b4299d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.594454 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "786b442d-b215-4707-b9b5-62f167b4299d" (UID: "786b442d-b215-4707-b9b5-62f167b4299d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.594616 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-config" (OuterVolumeSpecName: "config") pod "786b442d-b215-4707-b9b5-62f167b4299d" (UID: "786b442d-b215-4707-b9b5-62f167b4299d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.594784 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "786b442d-b215-4707-b9b5-62f167b4299d" (UID: "786b442d-b215-4707-b9b5-62f167b4299d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.616527 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.616559 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.616570 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b5thc\" (UniqueName: \"kubernetes.io/projected/786b442d-b215-4707-b9b5-62f167b4299d-kube-api-access-b5thc\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.616582 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.616609 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/786b442d-b215-4707-b9b5-62f167b4299d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.722881 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b559c875f-bl7vh"] Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.730441 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b559c875f-bl7vh"] Dec 05 09:29:20 crc kubenswrapper[4815]: I1205 09:29:20.987365 4815 scope.go:117] "RemoveContainer" containerID="d40f4852bd467be3c730273e865e2158952ef153741e7255587aea05fc2cad13" Dec 05 09:29:20 crc kubenswrapper[4815]: E1205 09:29:20.997730 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24@sha256:6b929971283d69f485a7d3e449fb5a3dd65d5a4de585c73419e776821d00062c" Dec 05 09:29:20 crc kubenswrapper[4815]: E1205 09:29:20.998078 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24@sha256:6b929971283d69f485a7d3e449fb5a3dd65d5a4de585c73419e776821d00062c,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5kgmv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(501b099d-0ffb-4c12-9809-15107cd1431a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 05 09:29:20 crc kubenswrapper[4815]: E1205 09:29:20.999376 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"]" pod="openstack/ceilometer-0" podUID="501b099d-0ffb-4c12-9809-15107cd1431a" Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.150590 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.164263 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.396199 4815 generic.go:334] "Generic (PLEG): container finished" podID="c66190b5-87b0-440b-9894-1bffb2cfca3f" containerID="ac13c63149b7164d533764020d964e0736c593645c983b6e58b63ac587517a86" exitCode=0 Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.396269 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ss9wk" event={"ID":"c66190b5-87b0-440b-9894-1bffb2cfca3f","Type":"ContainerDied","Data":"ac13c63149b7164d533764020d964e0736c593645c983b6e58b63ac587517a86"} Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.398041 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-588897d4d8-8nmk9" event={"ID":"d7266a34-7fa1-442b-9e83-e66665a5cd12","Type":"ContainerStarted","Data":"7acfdb3c0f65b6956c2f26207ba4a64728f5f5bbcc0f96acf58df59f707a0b00"} Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.399696 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.399729 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.407477 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-588897d4d8-8nmk9" podUID="d7266a34-7fa1-442b-9e83-e66665a5cd12" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.141:9311/healthcheck\": dial tcp 10.217.0.141:9311: connect: connection refused" Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.409215 4815 generic.go:334] "Generic (PLEG): container finished" podID="a2ea4b20-281a-43e1-b6af-c2c209b10e9d" containerID="d09086a87a983214a6a3881c0aa175532e712f484214c0057dc7f8fea016f095" exitCode=0 Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.409456 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dgjpz" event={"ID":"a2ea4b20-281a-43e1-b6af-c2c209b10e9d","Type":"ContainerDied","Data":"d09086a87a983214a6a3881c0aa175532e712f484214c0057dc7f8fea016f095"} Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.409673 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="501b099d-0ffb-4c12-9809-15107cd1431a" containerName="ceilometer-notification-agent" containerID="cri-o://7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792" gracePeriod=30 Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.409705 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="501b099d-0ffb-4c12-9809-15107cd1431a" containerName="sg-core" containerID="cri-o://19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d" gracePeriod=30 Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.430441 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="786b442d-b215-4707-b9b5-62f167b4299d" path="/var/lib/kubelet/pods/786b442d-b215-4707-b9b5-62f167b4299d/volumes" Dec 05 09:29:21 crc kubenswrapper[4815]: I1205 09:29:21.461693 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-588897d4d8-8nmk9" podStartSLOduration=18.461675399 podStartE2EDuration="18.461675399s" podCreationTimestamp="2025-12-05 09:29:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:21.458356467 +0000 UTC m=+1360.336963304" watchObservedRunningTime="2025-12-05 09:29:21.461675399 +0000 UTC m=+1360.340282236" Dec 05 09:29:22 crc kubenswrapper[4815]: I1205 09:29:22.658891 4815 generic.go:334] "Generic (PLEG): container finished" podID="c832b179-b3a7-45d6-a360-11e05100d1f2" containerID="c86c9131a647365e140b7d211616e5b78f71e312cf85293d078ba9093f0e52da" exitCode=0 Dec 05 09:29:22 crc kubenswrapper[4815]: I1205 09:29:22.658980 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-96bs8" event={"ID":"c832b179-b3a7-45d6-a360-11e05100d1f2","Type":"ContainerDied","Data":"c86c9131a647365e140b7d211616e5b78f71e312cf85293d078ba9093f0e52da"} Dec 05 09:29:22 crc kubenswrapper[4815]: I1205 09:29:22.669232 4815 generic.go:334] "Generic (PLEG): container finished" podID="501b099d-0ffb-4c12-9809-15107cd1431a" containerID="19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d" exitCode=2 Dec 05 09:29:22 crc kubenswrapper[4815]: I1205 09:29:22.669463 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501b099d-0ffb-4c12-9809-15107cd1431a","Type":"ContainerDied","Data":"19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d"} Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.192890 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ss9wk" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.358435 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.395539 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-combined-ca-bundle\") pod \"c66190b5-87b0-440b-9894-1bffb2cfca3f\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.395593 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4cftk\" (UniqueName: \"kubernetes.io/projected/c66190b5-87b0-440b-9894-1bffb2cfca3f-kube-api-access-4cftk\") pod \"c66190b5-87b0-440b-9894-1bffb2cfca3f\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.395634 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c66190b5-87b0-440b-9894-1bffb2cfca3f-logs\") pod \"c66190b5-87b0-440b-9894-1bffb2cfca3f\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.395711 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-scripts\") pod \"c66190b5-87b0-440b-9894-1bffb2cfca3f\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.395733 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-config-data\") pod \"c66190b5-87b0-440b-9894-1bffb2cfca3f\" (UID: \"c66190b5-87b0-440b-9894-1bffb2cfca3f\") " Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.396870 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c66190b5-87b0-440b-9894-1bffb2cfca3f-logs" (OuterVolumeSpecName: "logs") pod "c66190b5-87b0-440b-9894-1bffb2cfca3f" (UID: "c66190b5-87b0-440b-9894-1bffb2cfca3f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.403655 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c66190b5-87b0-440b-9894-1bffb2cfca3f-kube-api-access-4cftk" (OuterVolumeSpecName: "kube-api-access-4cftk") pod "c66190b5-87b0-440b-9894-1bffb2cfca3f" (UID: "c66190b5-87b0-440b-9894-1bffb2cfca3f"). InnerVolumeSpecName "kube-api-access-4cftk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.420747 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-scripts" (OuterVolumeSpecName: "scripts") pod "c66190b5-87b0-440b-9894-1bffb2cfca3f" (UID: "c66190b5-87b0-440b-9894-1bffb2cfca3f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.423773 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c66190b5-87b0-440b-9894-1bffb2cfca3f" (UID: "c66190b5-87b0-440b-9894-1bffb2cfca3f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.434930 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-config-data" (OuterVolumeSpecName: "config-data") pod "c66190b5-87b0-440b-9894-1bffb2cfca3f" (UID: "c66190b5-87b0-440b-9894-1bffb2cfca3f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.497640 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-config\") pod \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.498185 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-combined-ca-bundle\") pod \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.498470 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npqd8\" (UniqueName: \"kubernetes.io/projected/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-kube-api-access-npqd8\") pod \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\" (UID: \"a2ea4b20-281a-43e1-b6af-c2c209b10e9d\") " Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.500640 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.500784 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4cftk\" (UniqueName: \"kubernetes.io/projected/c66190b5-87b0-440b-9894-1bffb2cfca3f-kube-api-access-4cftk\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.500863 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c66190b5-87b0-440b-9894-1bffb2cfca3f-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.500937 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.501023 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c66190b5-87b0-440b-9894-1bffb2cfca3f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.502140 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-kube-api-access-npqd8" (OuterVolumeSpecName: "kube-api-access-npqd8") pod "a2ea4b20-281a-43e1-b6af-c2c209b10e9d" (UID: "a2ea4b20-281a-43e1-b6af-c2c209b10e9d"). InnerVolumeSpecName "kube-api-access-npqd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.521149 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a2ea4b20-281a-43e1-b6af-c2c209b10e9d" (UID: "a2ea4b20-281a-43e1-b6af-c2c209b10e9d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.528056 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-config" (OuterVolumeSpecName: "config") pod "a2ea4b20-281a-43e1-b6af-c2c209b10e9d" (UID: "a2ea4b20-281a-43e1-b6af-c2c209b10e9d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.601766 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npqd8\" (UniqueName: \"kubernetes.io/projected/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-kube-api-access-npqd8\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.602108 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.602179 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2ea4b20-281a-43e1-b6af-c2c209b10e9d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.707366 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-ss9wk" event={"ID":"c66190b5-87b0-440b-9894-1bffb2cfca3f","Type":"ContainerDied","Data":"b19745a03816abf498bc255d946b1d525a681a6224e1fec8dbccceb3214716b1"} Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.707412 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b19745a03816abf498bc255d946b1d525a681a6224e1fec8dbccceb3214716b1" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.707481 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-ss9wk" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.728699 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-dgjpz" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.729743 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-dgjpz" event={"ID":"a2ea4b20-281a-43e1-b6af-c2c209b10e9d","Type":"ContainerDied","Data":"3bcb739c62ee00862983222f7bb32b4fc2a7ea16f98dd9b0159baa4726e19f8d"} Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.729791 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bcb739c62ee00862983222f7bb32b4fc2a7ea16f98dd9b0159baa4726e19f8d" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.786159 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-549777fc4b-mjkbk"] Dec 05 09:29:23 crc kubenswrapper[4815]: E1205 09:29:23.810967 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b442d-b215-4707-b9b5-62f167b4299d" containerName="dnsmasq-dns" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.811204 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b442d-b215-4707-b9b5-62f167b4299d" containerName="dnsmasq-dns" Dec 05 09:29:23 crc kubenswrapper[4815]: E1205 09:29:23.811308 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="786b442d-b215-4707-b9b5-62f167b4299d" containerName="init" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.811370 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="786b442d-b215-4707-b9b5-62f167b4299d" containerName="init" Dec 05 09:29:23 crc kubenswrapper[4815]: E1205 09:29:23.811439 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c66190b5-87b0-440b-9894-1bffb2cfca3f" containerName="placement-db-sync" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.811523 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c66190b5-87b0-440b-9894-1bffb2cfca3f" containerName="placement-db-sync" Dec 05 09:29:23 crc kubenswrapper[4815]: E1205 09:29:23.811609 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2ea4b20-281a-43e1-b6af-c2c209b10e9d" containerName="neutron-db-sync" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.811685 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2ea4b20-281a-43e1-b6af-c2c209b10e9d" containerName="neutron-db-sync" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.811969 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="786b442d-b215-4707-b9b5-62f167b4299d" containerName="dnsmasq-dns" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.813109 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c66190b5-87b0-440b-9894-1bffb2cfca3f" containerName="placement-db-sync" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.813213 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2ea4b20-281a-43e1-b6af-c2c209b10e9d" containerName="neutron-db-sync" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.814344 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-549777fc4b-mjkbk"] Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.814532 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.821836 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.822264 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.822481 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.822756 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-dgw9g" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.829022 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.914135 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75817271-e950-4196-bc2c-b32ac4dad684-logs\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.922634 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-scripts\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.922769 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-internal-tls-certs\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.922799 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-public-tls-certs\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.922845 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfkgd\" (UniqueName: \"kubernetes.io/projected/75817271-e950-4196-bc2c-b32ac4dad684-kube-api-access-zfkgd\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.922929 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-combined-ca-bundle\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:23 crc kubenswrapper[4815]: I1205 09:29:23.922994 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-config-data\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.024842 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75817271-e950-4196-bc2c-b32ac4dad684-logs\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.024894 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-scripts\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.024956 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-internal-tls-certs\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.024988 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-public-tls-certs\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.025021 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfkgd\" (UniqueName: \"kubernetes.io/projected/75817271-e950-4196-bc2c-b32ac4dad684-kube-api-access-zfkgd\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.025082 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-combined-ca-bundle\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.025155 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-config-data\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.037281 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-internal-tls-certs\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.037641 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/75817271-e950-4196-bc2c-b32ac4dad684-logs\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.038367 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-combined-ca-bundle\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.039034 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-public-tls-certs\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.047800 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-scripts\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.055093 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-58b499c84f-xm6sx"] Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.056835 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.088710 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-58b499c84f-xm6sx"] Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.095776 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/75817271-e950-4196-bc2c-b32ac4dad684-config-data\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.108023 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfkgd\" (UniqueName: \"kubernetes.io/projected/75817271-e950-4196-bc2c-b32ac4dad684-kube-api-access-zfkgd\") pod \"placement-549777fc4b-mjkbk\" (UID: \"75817271-e950-4196-bc2c-b32ac4dad684\") " pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.126621 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-config\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.126700 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd5dq\" (UniqueName: \"kubernetes.io/projected/8da080f9-a053-474c-995b-7931442df050-kube-api-access-jd5dq\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.126747 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-nb\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.126823 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-sb\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.126856 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-dns-svc\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.135346 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6bdb56b95d-k5pvr"] Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.137109 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.147328 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.147762 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.151780 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-cwk9h" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.152018 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.168641 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6bdb56b95d-k5pvr"] Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.181254 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.228144 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-dns-svc\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.228221 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-combined-ca-bundle\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.228304 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8sx7l\" (UniqueName: \"kubernetes.io/projected/7126486e-29e9-487c-b603-7bf23251591b-kube-api-access-8sx7l\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.228336 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-config\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.228365 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd5dq\" (UniqueName: \"kubernetes.io/projected/8da080f9-a053-474c-995b-7931442df050-kube-api-access-jd5dq\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.228396 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-config\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.228422 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-nb\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.228457 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-ovndb-tls-certs\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.228514 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-httpd-config\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.228561 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-sb\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.229005 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-dns-svc\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.229457 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-sb\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.229542 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-config\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.234795 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-nb\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.256255 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd5dq\" (UniqueName: \"kubernetes.io/projected/8da080f9-a053-474c-995b-7931442df050-kube-api-access-jd5dq\") pod \"dnsmasq-dns-58b499c84f-xm6sx\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.335509 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-combined-ca-bundle\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.335593 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8sx7l\" (UniqueName: \"kubernetes.io/projected/7126486e-29e9-487c-b603-7bf23251591b-kube-api-access-8sx7l\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.335623 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-config\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.335657 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-ovndb-tls-certs\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.335680 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-httpd-config\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.352439 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-combined-ca-bundle\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.356465 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-ovndb-tls-certs\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.358088 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-config\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.358674 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8sx7l\" (UniqueName: \"kubernetes.io/projected/7126486e-29e9-487c-b603-7bf23251591b-kube-api-access-8sx7l\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.359778 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-httpd-config\") pod \"neutron-6bdb56b95d-k5pvr\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.489993 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-96bs8" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.511222 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.538698 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-config-data\") pod \"c832b179-b3a7-45d6-a360-11e05100d1f2\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.538961 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-scripts\") pod \"c832b179-b3a7-45d6-a360-11e05100d1f2\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.539065 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-db-sync-config-data\") pod \"c832b179-b3a7-45d6-a360-11e05100d1f2\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.539180 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7vzd\" (UniqueName: \"kubernetes.io/projected/c832b179-b3a7-45d6-a360-11e05100d1f2-kube-api-access-m7vzd\") pod \"c832b179-b3a7-45d6-a360-11e05100d1f2\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.539281 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c832b179-b3a7-45d6-a360-11e05100d1f2-etc-machine-id\") pod \"c832b179-b3a7-45d6-a360-11e05100d1f2\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.539383 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-combined-ca-bundle\") pod \"c832b179-b3a7-45d6-a360-11e05100d1f2\" (UID: \"c832b179-b3a7-45d6-a360-11e05100d1f2\") " Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.545778 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c832b179-b3a7-45d6-a360-11e05100d1f2-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c832b179-b3a7-45d6-a360-11e05100d1f2" (UID: "c832b179-b3a7-45d6-a360-11e05100d1f2"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.560328 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-scripts" (OuterVolumeSpecName: "scripts") pod "c832b179-b3a7-45d6-a360-11e05100d1f2" (UID: "c832b179-b3a7-45d6-a360-11e05100d1f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.569817 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "c832b179-b3a7-45d6-a360-11e05100d1f2" (UID: "c832b179-b3a7-45d6-a360-11e05100d1f2"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.572917 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c832b179-b3a7-45d6-a360-11e05100d1f2-kube-api-access-m7vzd" (OuterVolumeSpecName: "kube-api-access-m7vzd") pod "c832b179-b3a7-45d6-a360-11e05100d1f2" (UID: "c832b179-b3a7-45d6-a360-11e05100d1f2"). InnerVolumeSpecName "kube-api-access-m7vzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.625006 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.638394 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c832b179-b3a7-45d6-a360-11e05100d1f2" (UID: "c832b179-b3a7-45d6-a360-11e05100d1f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.643141 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.643190 4815 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.643203 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7vzd\" (UniqueName: \"kubernetes.io/projected/c832b179-b3a7-45d6-a360-11e05100d1f2-kube-api-access-m7vzd\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.643214 4815 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c832b179-b3a7-45d6-a360-11e05100d1f2-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.643222 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.652709 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-config-data" (OuterVolumeSpecName: "config-data") pod "c832b179-b3a7-45d6-a360-11e05100d1f2" (UID: "c832b179-b3a7-45d6-a360-11e05100d1f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.757860 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c832b179-b3a7-45d6-a360-11e05100d1f2-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.800213 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-96bs8" event={"ID":"c832b179-b3a7-45d6-a360-11e05100d1f2","Type":"ContainerDied","Data":"47b3b5497a244b810b78019cb0f3cbb0180745c82ef73c332c8adbfd2ff7de13"} Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.800262 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47b3b5497a244b810b78019cb0f3cbb0180745c82ef73c332c8adbfd2ff7de13" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.800358 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-96bs8" Dec 05 09:29:24 crc kubenswrapper[4815]: I1205 09:29:24.857226 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-549777fc4b-mjkbk"] Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.070594 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:29:25 crc kubenswrapper[4815]: E1205 09:29:25.071294 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c832b179-b3a7-45d6-a360-11e05100d1f2" containerName="cinder-db-sync" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.071311 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c832b179-b3a7-45d6-a360-11e05100d1f2" containerName="cinder-db-sync" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.071573 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c832b179-b3a7-45d6-a360-11e05100d1f2" containerName="cinder-db-sync" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.072681 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.084998 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.085193 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.085411 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-5z58g" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.085567 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.107672 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.166668 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.166704 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.166748 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.166887 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x26b7\" (UniqueName: \"kubernetes.io/projected/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-kube-api-access-x26b7\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.166995 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.167020 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-scripts\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.170577 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58b499c84f-xm6sx"] Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.233349 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-799db77f-npk6z"] Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.234669 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.268024 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.268055 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.268088 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.268125 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-dns-svc\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.268180 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x26b7\" (UniqueName: \"kubernetes.io/projected/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-kube-api-access-x26b7\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.268198 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-config\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.268222 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.268240 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-scripts\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.268259 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-nb\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.268298 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-sb\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.268317 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27nn4\" (UniqueName: \"kubernetes.io/projected/f042ee2e-f5c8-46af-b494-523064e9b7cd-kube-api-access-27nn4\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.269463 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.270556 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-799db77f-npk6z"] Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.289538 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.292474 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.293004 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.293365 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-scripts\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.330606 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x26b7\" (UniqueName: \"kubernetes.io/projected/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-kube-api-access-x26b7\") pod \"cinder-scheduler-0\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.342650 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58b499c84f-xm6sx"] Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.370635 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-config\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.370728 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-nb\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.370771 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-sb\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.370797 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27nn4\" (UniqueName: \"kubernetes.io/projected/f042ee2e-f5c8-46af-b494-523064e9b7cd-kube-api-access-27nn4\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.370893 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-dns-svc\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.371928 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-dns-svc\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.372478 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-sb\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.372925 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-nb\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.373099 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-config\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.405392 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.407211 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.419181 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.419566 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27nn4\" (UniqueName: \"kubernetes.io/projected/f042ee2e-f5c8-46af-b494-523064e9b7cd-kube-api-access-27nn4\") pod \"dnsmasq-dns-799db77f-npk6z\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.475164 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-scripts\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.475539 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.475658 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d38833-a5a9-4961-9ce8-7722b58624a3-logs\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.475694 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data-custom\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.475742 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.475812 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d38833-a5a9-4961-9ce8-7722b58624a3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.475840 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkffl\" (UniqueName: \"kubernetes.io/projected/b6d38833-a5a9-4961-9ce8-7722b58624a3-kube-api-access-zkffl\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.480862 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.481918 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.567919 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.577714 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d38833-a5a9-4961-9ce8-7722b58624a3-logs\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.577772 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data-custom\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.577821 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.577870 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d38833-a5a9-4961-9ce8-7722b58624a3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.577900 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkffl\" (UniqueName: \"kubernetes.io/projected/b6d38833-a5a9-4961-9ce8-7722b58624a3-kube-api-access-zkffl\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.577947 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-scripts\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.577964 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.578274 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d38833-a5a9-4961-9ce8-7722b58624a3-logs\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.578347 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d38833-a5a9-4961-9ce8-7722b58624a3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.590063 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data-custom\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.596984 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-scripts\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.598391 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.598950 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.610535 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkffl\" (UniqueName: \"kubernetes.io/projected/b6d38833-a5a9-4961-9ce8-7722b58624a3-kube-api-access-zkffl\") pod \"cinder-api-0\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.676005 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6bdb56b95d-k5pvr"] Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.818281 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.836723 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" event={"ID":"8da080f9-a053-474c-995b-7931442df050","Type":"ContainerStarted","Data":"c962db2c9757c1b5e30a5f892dd2391e3c4dcf5e334024a012554c22eef7c742"} Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.839561 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-549777fc4b-mjkbk" event={"ID":"75817271-e950-4196-bc2c-b32ac4dad684","Type":"ContainerStarted","Data":"7b0509546c94e61e4e403bfb0a1727520efeab1148c09ab3d5775563f25e4fc4"} Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.839599 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-549777fc4b-mjkbk" event={"ID":"75817271-e950-4196-bc2c-b32ac4dad684","Type":"ContainerStarted","Data":"1bff512ad01b435a313af47a2b4a4ce92d2e33caad6ee5e9ee5a9ec675485596"} Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.839609 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-549777fc4b-mjkbk" event={"ID":"75817271-e950-4196-bc2c-b32ac4dad684","Type":"ContainerStarted","Data":"ba9734899f9e56f039a86f6e777b70ab573cebe489023fa8aca6dc3cb91989b4"} Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.841222 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.841270 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.842471 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdb56b95d-k5pvr" event={"ID":"7126486e-29e9-487c-b603-7bf23251591b","Type":"ContainerStarted","Data":"1bf4df7c20ea4af99bcb76413a080170c1d6096dca3b00ab23731d5711af0964"} Dec 05 09:29:25 crc kubenswrapper[4815]: I1205 09:29:25.885512 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-549777fc4b-mjkbk" podStartSLOduration=2.885495917 podStartE2EDuration="2.885495917s" podCreationTimestamp="2025-12-05 09:29:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:25.880320354 +0000 UTC m=+1364.758927191" watchObservedRunningTime="2025-12-05 09:29:25.885495917 +0000 UTC m=+1364.764102754" Dec 05 09:29:26 crc kubenswrapper[4815]: I1205 09:29:26.029970 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:29:26 crc kubenswrapper[4815]: I1205 09:29:26.370293 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-799db77f-npk6z"] Dec 05 09:29:26 crc kubenswrapper[4815]: I1205 09:29:26.762637 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 09:29:26 crc kubenswrapper[4815]: W1205 09:29:26.778283 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb6d38833_a5a9_4961_9ce8_7722b58624a3.slice/crio-265021055cef23d3ac00f988b6cf8c28312e8cdb4423cf804b7a29792d8841b8 WatchSource:0}: Error finding container 265021055cef23d3ac00f988b6cf8c28312e8cdb4423cf804b7a29792d8841b8: Status 404 returned error can't find the container with id 265021055cef23d3ac00f988b6cf8c28312e8cdb4423cf804b7a29792d8841b8 Dec 05 09:29:26 crc kubenswrapper[4815]: I1205 09:29:26.865037 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891","Type":"ContainerStarted","Data":"ae2300cddaf4a952d8ee6ec0ff6cd2505a6315dea681093657ffb4ad93db09ce"} Dec 05 09:29:26 crc kubenswrapper[4815]: I1205 09:29:26.866993 4815 generic.go:334] "Generic (PLEG): container finished" podID="8da080f9-a053-474c-995b-7931442df050" containerID="c6246228200b3b2905e499e7f4036a9bc4d67d99e75e9fca01aa101db2bdcd8f" exitCode=0 Dec 05 09:29:26 crc kubenswrapper[4815]: I1205 09:29:26.867038 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" event={"ID":"8da080f9-a053-474c-995b-7931442df050","Type":"ContainerDied","Data":"c6246228200b3b2905e499e7f4036a9bc4d67d99e75e9fca01aa101db2bdcd8f"} Dec 05 09:29:26 crc kubenswrapper[4815]: I1205 09:29:26.869609 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdb56b95d-k5pvr" event={"ID":"7126486e-29e9-487c-b603-7bf23251591b","Type":"ContainerStarted","Data":"697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab"} Dec 05 09:29:26 crc kubenswrapper[4815]: I1205 09:29:26.875954 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-799db77f-npk6z" event={"ID":"f042ee2e-f5c8-46af-b494-523064e9b7cd","Type":"ContainerStarted","Data":"c56856aadf4f7eeecc75ff86811d67d8b26b63bdf60986a22e9c80f88bf5e164"} Dec 05 09:29:26 crc kubenswrapper[4815]: I1205 09:29:26.878527 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d38833-a5a9-4961-9ce8-7722b58624a3","Type":"ContainerStarted","Data":"265021055cef23d3ac00f988b6cf8c28312e8cdb4423cf804b7a29792d8841b8"} Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.401227 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.483009 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.540907 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-config\") pod \"8da080f9-a053-474c-995b-7931442df050\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.540989 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-dns-svc\") pod \"8da080f9-a053-474c-995b-7931442df050\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.541023 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jd5dq\" (UniqueName: \"kubernetes.io/projected/8da080f9-a053-474c-995b-7931442df050-kube-api-access-jd5dq\") pod \"8da080f9-a053-474c-995b-7931442df050\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.541061 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-nb\") pod \"8da080f9-a053-474c-995b-7931442df050\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.541143 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-sb\") pod \"8da080f9-a053-474c-995b-7931442df050\" (UID: \"8da080f9-a053-474c-995b-7931442df050\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.549651 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8da080f9-a053-474c-995b-7931442df050-kube-api-access-jd5dq" (OuterVolumeSpecName: "kube-api-access-jd5dq") pod "8da080f9-a053-474c-995b-7931442df050" (UID: "8da080f9-a053-474c-995b-7931442df050"). InnerVolumeSpecName "kube-api-access-jd5dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.566714 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-config" (OuterVolumeSpecName: "config") pod "8da080f9-a053-474c-995b-7931442df050" (UID: "8da080f9-a053-474c-995b-7931442df050"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.578771 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8da080f9-a053-474c-995b-7931442df050" (UID: "8da080f9-a053-474c-995b-7931442df050"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.579057 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8da080f9-a053-474c-995b-7931442df050" (UID: "8da080f9-a053-474c-995b-7931442df050"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.618930 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8da080f9-a053-474c-995b-7931442df050" (UID: "8da080f9-a053-474c-995b-7931442df050"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.643592 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.643627 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.643640 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jd5dq\" (UniqueName: \"kubernetes.io/projected/8da080f9-a053-474c-995b-7931442df050-kube-api-access-jd5dq\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.643652 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.643663 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8da080f9-a053-474c-995b-7931442df050-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.784402 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.845628 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-run-httpd\") pod \"501b099d-0ffb-4c12-9809-15107cd1431a\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.845674 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-scripts\") pod \"501b099d-0ffb-4c12-9809-15107cd1431a\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.845786 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kgmv\" (UniqueName: \"kubernetes.io/projected/501b099d-0ffb-4c12-9809-15107cd1431a-kube-api-access-5kgmv\") pod \"501b099d-0ffb-4c12-9809-15107cd1431a\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.845948 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-log-httpd\") pod \"501b099d-0ffb-4c12-9809-15107cd1431a\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.845975 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-config-data\") pod \"501b099d-0ffb-4c12-9809-15107cd1431a\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.845997 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-sg-core-conf-yaml\") pod \"501b099d-0ffb-4c12-9809-15107cd1431a\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.846017 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-combined-ca-bundle\") pod \"501b099d-0ffb-4c12-9809-15107cd1431a\" (UID: \"501b099d-0ffb-4c12-9809-15107cd1431a\") " Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.849829 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "501b099d-0ffb-4c12-9809-15107cd1431a" (UID: "501b099d-0ffb-4c12-9809-15107cd1431a"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.850000 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "501b099d-0ffb-4c12-9809-15107cd1431a" (UID: "501b099d-0ffb-4c12-9809-15107cd1431a"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.863011 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/501b099d-0ffb-4c12-9809-15107cd1431a-kube-api-access-5kgmv" (OuterVolumeSpecName: "kube-api-access-5kgmv") pod "501b099d-0ffb-4c12-9809-15107cd1431a" (UID: "501b099d-0ffb-4c12-9809-15107cd1431a"). InnerVolumeSpecName "kube-api-access-5kgmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.865610 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-scripts" (OuterVolumeSpecName: "scripts") pod "501b099d-0ffb-4c12-9809-15107cd1431a" (UID: "501b099d-0ffb-4c12-9809-15107cd1431a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.876166 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-config-data" (OuterVolumeSpecName: "config-data") pod "501b099d-0ffb-4c12-9809-15107cd1431a" (UID: "501b099d-0ffb-4c12-9809-15107cd1431a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.886682 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "501b099d-0ffb-4c12-9809-15107cd1431a" (UID: "501b099d-0ffb-4c12-9809-15107cd1431a"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.891099 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "501b099d-0ffb-4c12-9809-15107cd1431a" (UID: "501b099d-0ffb-4c12-9809-15107cd1431a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.893711 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdb56b95d-k5pvr" event={"ID":"7126486e-29e9-487c-b603-7bf23251591b","Type":"ContainerStarted","Data":"9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35"} Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.894929 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.902206 4815 generic.go:334] "Generic (PLEG): container finished" podID="501b099d-0ffb-4c12-9809-15107cd1431a" containerID="7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792" exitCode=0 Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.902282 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501b099d-0ffb-4c12-9809-15107cd1431a","Type":"ContainerDied","Data":"7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792"} Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.902310 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"501b099d-0ffb-4c12-9809-15107cd1431a","Type":"ContainerDied","Data":"3eecbebffc2f5140c795577bcc53a3f686652af35b0a1f83ffeb99fce0a3a55c"} Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.902346 4815 scope.go:117] "RemoveContainer" containerID="19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.902508 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.920505 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6bdb56b95d-k5pvr" podStartSLOduration=3.9204665629999997 podStartE2EDuration="3.920466563s" podCreationTimestamp="2025-12-05 09:29:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:27.914201902 +0000 UTC m=+1366.792808739" watchObservedRunningTime="2025-12-05 09:29:27.920466563 +0000 UTC m=+1366.799073400" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.920882 4815 generic.go:334] "Generic (PLEG): container finished" podID="f042ee2e-f5c8-46af-b494-523064e9b7cd" containerID="7518a815d82fdf4bf9a6d7bc1fce3eb51821b37429b22d390bc1f754166e1718" exitCode=0 Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.920953 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-799db77f-npk6z" event={"ID":"f042ee2e-f5c8-46af-b494-523064e9b7cd","Type":"ContainerDied","Data":"7518a815d82fdf4bf9a6d7bc1fce3eb51821b37429b22d390bc1f754166e1718"} Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.947533 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kgmv\" (UniqueName: \"kubernetes.io/projected/501b099d-0ffb-4c12-9809-15107cd1431a-kube-api-access-5kgmv\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.947881 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.947997 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.948056 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.948111 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.948162 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/501b099d-0ffb-4c12-9809-15107cd1431a-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.948235 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/501b099d-0ffb-4c12-9809-15107cd1431a-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.972650 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" Dec 05 09:29:27 crc kubenswrapper[4815]: I1205 09:29:27.973280 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-58b499c84f-xm6sx" event={"ID":"8da080f9-a053-474c-995b-7931442df050","Type":"ContainerDied","Data":"c962db2c9757c1b5e30a5f892dd2391e3c4dcf5e334024a012554c22eef7c742"} Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.054563 4815 scope.go:117] "RemoveContainer" containerID="7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.059775 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.096735 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.163213 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:29:28 crc kubenswrapper[4815]: E1205 09:29:28.163845 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8da080f9-a053-474c-995b-7931442df050" containerName="init" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.163866 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8da080f9-a053-474c-995b-7931442df050" containerName="init" Dec 05 09:29:28 crc kubenswrapper[4815]: E1205 09:29:28.163879 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501b099d-0ffb-4c12-9809-15107cd1431a" containerName="ceilometer-notification-agent" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.163885 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="501b099d-0ffb-4c12-9809-15107cd1431a" containerName="ceilometer-notification-agent" Dec 05 09:29:28 crc kubenswrapper[4815]: E1205 09:29:28.163897 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="501b099d-0ffb-4c12-9809-15107cd1431a" containerName="sg-core" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.163904 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="501b099d-0ffb-4c12-9809-15107cd1431a" containerName="sg-core" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.164622 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="501b099d-0ffb-4c12-9809-15107cd1431a" containerName="ceilometer-notification-agent" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.164690 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="501b099d-0ffb-4c12-9809-15107cd1431a" containerName="sg-core" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.164706 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8da080f9-a053-474c-995b-7931442df050" containerName="init" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.172510 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.188750 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.189055 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.194222 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.232957 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-58b499c84f-xm6sx"] Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.244569 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-58b499c84f-xm6sx"] Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.260914 4815 scope.go:117] "RemoveContainer" containerID="19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d" Dec 05 09:29:28 crc kubenswrapper[4815]: E1205 09:29:28.261399 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d\": container with ID starting with 19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d not found: ID does not exist" containerID="19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.261428 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d"} err="failed to get container status \"19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d\": rpc error: code = NotFound desc = could not find container \"19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d\": container with ID starting with 19596c7bfe675bcf6a8abff7f5b99f9cee88bdc39a2551b2c74b7c975884898d not found: ID does not exist" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.261449 4815 scope.go:117] "RemoveContainer" containerID="7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792" Dec 05 09:29:28 crc kubenswrapper[4815]: E1205 09:29:28.263022 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792\": container with ID starting with 7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792 not found: ID does not exist" containerID="7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.263047 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792"} err="failed to get container status \"7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792\": rpc error: code = NotFound desc = could not find container \"7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792\": container with ID starting with 7fe12e7f4087fdb99003efffce8b18a94456095f0928d7883629f91ae2557792 not found: ID does not exist" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.263063 4815 scope.go:117] "RemoveContainer" containerID="c6246228200b3b2905e499e7f4036a9bc4d67d99e75e9fca01aa101db2bdcd8f" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.269396 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-scripts\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.269465 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.269502 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.269559 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-config-data\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.269577 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-run-httpd\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.269642 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-log-httpd\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.269664 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6f7m\" (UniqueName: \"kubernetes.io/projected/158b54fd-bab0-4163-8334-2a542378852e-kube-api-access-p6f7m\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.372959 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-scripts\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.373028 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.373049 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.373094 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-config-data\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.373110 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-run-httpd\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.373173 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-log-httpd\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.373196 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6f7m\" (UniqueName: \"kubernetes.io/projected/158b54fd-bab0-4163-8334-2a542378852e-kube-api-access-p6f7m\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.379306 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-scripts\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.382677 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.396345 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.397914 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-log-httpd\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.399781 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.407591 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-run-httpd\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.408332 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-config-data\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.421008 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6f7m\" (UniqueName: \"kubernetes.io/projected/158b54fd-bab0-4163-8334-2a542378852e-kube-api-access-p6f7m\") pod \"ceilometer-0\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " pod="openstack/ceilometer-0" Dec 05 09:29:28 crc kubenswrapper[4815]: I1205 09:29:28.556611 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:29:29 crc kubenswrapper[4815]: I1205 09:29:29.000757 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-799db77f-npk6z" event={"ID":"f042ee2e-f5c8-46af-b494-523064e9b7cd","Type":"ContainerStarted","Data":"7f2eb5874d49f85f7ac50f51efe41dfd444da805af11a9738249a868e55988a6"} Dec 05 09:29:29 crc kubenswrapper[4815]: I1205 09:29:29.001018 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:29 crc kubenswrapper[4815]: I1205 09:29:29.008159 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d38833-a5a9-4961-9ce8-7722b58624a3","Type":"ContainerStarted","Data":"e5f13c2074adc75e0ab860daf53f09f877d7815781c669c2a11066e38a640d4f"} Dec 05 09:29:29 crc kubenswrapper[4815]: I1205 09:29:29.010034 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891","Type":"ContainerStarted","Data":"b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5"} Dec 05 09:29:29 crc kubenswrapper[4815]: I1205 09:29:29.027333 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-799db77f-npk6z" podStartSLOduration=4.027317585 podStartE2EDuration="4.027317585s" podCreationTimestamp="2025-12-05 09:29:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:29.023718426 +0000 UTC m=+1367.902325263" watchObservedRunningTime="2025-12-05 09:29:29.027317585 +0000 UTC m=+1367.905924422" Dec 05 09:29:29 crc kubenswrapper[4815]: I1205 09:29:29.209238 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:29:29 crc kubenswrapper[4815]: W1205 09:29:29.219444 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod158b54fd_bab0_4163_8334_2a542378852e.slice/crio-fd11f7591d0ef77e77394b1e130b9dcb7c89d1264ab02d5a13b6b5628ac52bca WatchSource:0}: Error finding container fd11f7591d0ef77e77394b1e130b9dcb7c89d1264ab02d5a13b6b5628ac52bca: Status 404 returned error can't find the container with id fd11f7591d0ef77e77394b1e130b9dcb7c89d1264ab02d5a13b6b5628ac52bca Dec 05 09:29:29 crc kubenswrapper[4815]: I1205 09:29:29.467010 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="501b099d-0ffb-4c12-9809-15107cd1431a" path="/var/lib/kubelet/pods/501b099d-0ffb-4c12-9809-15107cd1431a/volumes" Dec 05 09:29:29 crc kubenswrapper[4815]: I1205 09:29:29.468108 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8da080f9-a053-474c-995b-7931442df050" path="/var/lib/kubelet/pods/8da080f9-a053-474c-995b-7931442df050/volumes" Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.026137 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d38833-a5a9-4961-9ce8-7722b58624a3","Type":"ContainerStarted","Data":"d801d3dcbdb9a9aa7ecda716873f76db670edaa8093c20110c64ac69368920ff"} Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.026280 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b6d38833-a5a9-4961-9ce8-7722b58624a3" containerName="cinder-api-log" containerID="cri-o://e5f13c2074adc75e0ab860daf53f09f877d7815781c669c2a11066e38a640d4f" gracePeriod=30 Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.026613 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.026863 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="b6d38833-a5a9-4961-9ce8-7722b58624a3" containerName="cinder-api" containerID="cri-o://d801d3dcbdb9a9aa7ecda716873f76db670edaa8093c20110c64ac69368920ff" gracePeriod=30 Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.030289 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891","Type":"ContainerStarted","Data":"b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5"} Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.032246 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"158b54fd-bab0-4163-8334-2a542378852e","Type":"ContainerStarted","Data":"fd11f7591d0ef77e77394b1e130b9dcb7c89d1264ab02d5a13b6b5628ac52bca"} Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.052678 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.052663209 podStartE2EDuration="5.052663209s" podCreationTimestamp="2025-12-05 09:29:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:30.047409115 +0000 UTC m=+1368.926015942" watchObservedRunningTime="2025-12-05 09:29:30.052663209 +0000 UTC m=+1368.931270046" Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.078271 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.746720893 podStartE2EDuration="5.078247041s" podCreationTimestamp="2025-12-05 09:29:25 +0000 UTC" firstStartedPulling="2025-12-05 09:29:26.031075503 +0000 UTC m=+1364.909682340" lastFinishedPulling="2025-12-05 09:29:27.362601651 +0000 UTC m=+1366.241208488" observedRunningTime="2025-12-05 09:29:30.072708 +0000 UTC m=+1368.951314837" watchObservedRunningTime="2025-12-05 09:29:30.078247041 +0000 UTC m=+1368.956853878" Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.485641 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.657792 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-588897d4d8-8nmk9" Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.758720 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-cd8fb7c76-9g4zb"] Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.759243 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api-log" containerID="cri-o://45ef5ea6df8164df3bad60882166ae470ca705e908389d422fa6fc2185fe4ea0" gracePeriod=30 Dec 05 09:29:30 crc kubenswrapper[4815]: I1205 09:29:30.759720 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api" containerID="cri-o://fba393df4f5ab9c99d2166ead3e4b4b3180ea1a121161d1a36e462c3245f3a2a" gracePeriod=30 Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.041033 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-54c4865f77-tppk8"] Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.045345 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.049256 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.049419 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.072220 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54c4865f77-tppk8"] Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.075287 4815 generic.go:334] "Generic (PLEG): container finished" podID="b6d38833-a5a9-4961-9ce8-7722b58624a3" containerID="d801d3dcbdb9a9aa7ecda716873f76db670edaa8093c20110c64ac69368920ff" exitCode=0 Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.075320 4815 generic.go:334] "Generic (PLEG): container finished" podID="b6d38833-a5a9-4961-9ce8-7722b58624a3" containerID="e5f13c2074adc75e0ab860daf53f09f877d7815781c669c2a11066e38a640d4f" exitCode=143 Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.075356 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d38833-a5a9-4961-9ce8-7722b58624a3","Type":"ContainerDied","Data":"d801d3dcbdb9a9aa7ecda716873f76db670edaa8093c20110c64ac69368920ff"} Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.075442 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d38833-a5a9-4961-9ce8-7722b58624a3","Type":"ContainerDied","Data":"e5f13c2074adc75e0ab860daf53f09f877d7815781c669c2a11066e38a640d4f"} Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.078977 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"158b54fd-bab0-4163-8334-2a542378852e","Type":"ContainerStarted","Data":"fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3"} Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.104608 4815 generic.go:334] "Generic (PLEG): container finished" podID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerID="45ef5ea6df8164df3bad60882166ae470ca705e908389d422fa6fc2185fe4ea0" exitCode=143 Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.104830 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cd8fb7c76-9g4zb" event={"ID":"2acc91d0-1a89-4cef-bcbb-90bb61e1323f","Type":"ContainerDied","Data":"45ef5ea6df8164df3bad60882166ae470ca705e908389d422fa6fc2185fe4ea0"} Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.105960 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-ovndb-tls-certs\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.106239 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-public-tls-certs\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.106370 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc5g4\" (UniqueName: \"kubernetes.io/projected/fceefc10-15e5-4b68-9bb6-e786dde63941-kube-api-access-wc5g4\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.106507 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-config\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.106594 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-combined-ca-bundle\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.106698 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-httpd-config\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.106797 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-internal-tls-certs\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.209098 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-httpd-config\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.209555 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-internal-tls-certs\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.209599 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-ovndb-tls-certs\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.209823 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-public-tls-certs\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.209878 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc5g4\" (UniqueName: \"kubernetes.io/projected/fceefc10-15e5-4b68-9bb6-e786dde63941-kube-api-access-wc5g4\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.209913 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-config\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.209958 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-combined-ca-bundle\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.222356 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-httpd-config\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.222584 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-combined-ca-bundle\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.223172 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-public-tls-certs\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.224160 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-internal-tls-certs\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.224750 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-config\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.241111 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc5g4\" (UniqueName: \"kubernetes.io/projected/fceefc10-15e5-4b68-9bb6-e786dde63941-kube-api-access-wc5g4\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.286695 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/fceefc10-15e5-4b68-9bb6-e786dde63941-ovndb-tls-certs\") pod \"neutron-54c4865f77-tppk8\" (UID: \"fceefc10-15e5-4b68-9bb6-e786dde63941\") " pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.390603 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.422209 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.523168 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data\") pod \"b6d38833-a5a9-4961-9ce8-7722b58624a3\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.523235 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d38833-a5a9-4961-9ce8-7722b58624a3-logs\") pod \"b6d38833-a5a9-4961-9ce8-7722b58624a3\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.523308 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-combined-ca-bundle\") pod \"b6d38833-a5a9-4961-9ce8-7722b58624a3\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.523333 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d38833-a5a9-4961-9ce8-7722b58624a3-etc-machine-id\") pod \"b6d38833-a5a9-4961-9ce8-7722b58624a3\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.523346 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-scripts\") pod \"b6d38833-a5a9-4961-9ce8-7722b58624a3\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.523390 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkffl\" (UniqueName: \"kubernetes.io/projected/b6d38833-a5a9-4961-9ce8-7722b58624a3-kube-api-access-zkffl\") pod \"b6d38833-a5a9-4961-9ce8-7722b58624a3\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.523446 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data-custom\") pod \"b6d38833-a5a9-4961-9ce8-7722b58624a3\" (UID: \"b6d38833-a5a9-4961-9ce8-7722b58624a3\") " Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.531743 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6d38833-a5a9-4961-9ce8-7722b58624a3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "b6d38833-a5a9-4961-9ce8-7722b58624a3" (UID: "b6d38833-a5a9-4961-9ce8-7722b58624a3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.532641 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6d38833-a5a9-4961-9ce8-7722b58624a3-logs" (OuterVolumeSpecName: "logs") pod "b6d38833-a5a9-4961-9ce8-7722b58624a3" (UID: "b6d38833-a5a9-4961-9ce8-7722b58624a3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.552755 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "b6d38833-a5a9-4961-9ce8-7722b58624a3" (UID: "b6d38833-a5a9-4961-9ce8-7722b58624a3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.555693 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-scripts" (OuterVolumeSpecName: "scripts") pod "b6d38833-a5a9-4961-9ce8-7722b58624a3" (UID: "b6d38833-a5a9-4961-9ce8-7722b58624a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.575962 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6d38833-a5a9-4961-9ce8-7722b58624a3-kube-api-access-zkffl" (OuterVolumeSpecName: "kube-api-access-zkffl") pod "b6d38833-a5a9-4961-9ce8-7722b58624a3" (UID: "b6d38833-a5a9-4961-9ce8-7722b58624a3"). InnerVolumeSpecName "kube-api-access-zkffl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.631728 4815 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.631771 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d38833-a5a9-4961-9ce8-7722b58624a3-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.631784 4815 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d38833-a5a9-4961-9ce8-7722b58624a3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.631796 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.631808 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkffl\" (UniqueName: \"kubernetes.io/projected/b6d38833-a5a9-4961-9ce8-7722b58624a3-kube-api-access-zkffl\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.656100 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b6d38833-a5a9-4961-9ce8-7722b58624a3" (UID: "b6d38833-a5a9-4961-9ce8-7722b58624a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.675698 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data" (OuterVolumeSpecName: "config-data") pod "b6d38833-a5a9-4961-9ce8-7722b58624a3" (UID: "b6d38833-a5a9-4961-9ce8-7722b58624a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.734593 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:31 crc kubenswrapper[4815]: I1205 09:29:31.734615 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d38833-a5a9-4961-9ce8-7722b58624a3-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.113989 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"158b54fd-bab0-4163-8334-2a542378852e","Type":"ContainerStarted","Data":"66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6"} Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.117648 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.118763 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d38833-a5a9-4961-9ce8-7722b58624a3","Type":"ContainerDied","Data":"265021055cef23d3ac00f988b6cf8c28312e8cdb4423cf804b7a29792d8841b8"} Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.118842 4815 scope.go:117] "RemoveContainer" containerID="d801d3dcbdb9a9aa7ecda716873f76db670edaa8093c20110c64ac69368920ff" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.150938 4815 scope.go:117] "RemoveContainer" containerID="e5f13c2074adc75e0ab860daf53f09f877d7815781c669c2a11066e38a640d4f" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.184764 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-54c4865f77-tppk8"] Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.383630 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.484565 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.526081 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 05 09:29:32 crc kubenswrapper[4815]: E1205 09:29:32.534732 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d38833-a5a9-4961-9ce8-7722b58624a3" containerName="cinder-api-log" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.534771 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d38833-a5a9-4961-9ce8-7722b58624a3" containerName="cinder-api-log" Dec 05 09:29:32 crc kubenswrapper[4815]: E1205 09:29:32.534786 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6d38833-a5a9-4961-9ce8-7722b58624a3" containerName="cinder-api" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.534792 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6d38833-a5a9-4961-9ce8-7722b58624a3" containerName="cinder-api" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.535077 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d38833-a5a9-4961-9ce8-7722b58624a3" containerName="cinder-api-log" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.535096 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6d38833-a5a9-4961-9ce8-7722b58624a3" containerName="cinder-api" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.536115 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.545437 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.546474 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.546777 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.547633 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.573951 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.574043 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.574067 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-config-data\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.574085 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-scripts\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.574124 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.574160 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-config-data-custom\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.574228 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.574250 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-logs\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.574301 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2v7f\" (UniqueName: \"kubernetes.io/projected/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-kube-api-access-g2v7f\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.677448 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-config-data-custom\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.677543 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.677566 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-logs\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.677609 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2v7f\" (UniqueName: \"kubernetes.io/projected/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-kube-api-access-g2v7f\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.677640 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.677683 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.677706 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-config-data\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.677722 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-scripts\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.677745 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.678864 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-etc-machine-id\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.684766 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-config-data-custom\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.685032 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-logs\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.688970 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-public-tls-certs\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.689528 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.689958 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.694063 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-scripts\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.694497 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-config-data\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.723133 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2v7f\" (UniqueName: \"kubernetes.io/projected/b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05-kube-api-access-g2v7f\") pod \"cinder-api-0\" (UID: \"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05\") " pod="openstack/cinder-api-0" Dec 05 09:29:32 crc kubenswrapper[4815]: I1205 09:29:32.866418 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 05 09:29:33 crc kubenswrapper[4815]: I1205 09:29:33.156202 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"158b54fd-bab0-4163-8334-2a542378852e","Type":"ContainerStarted","Data":"13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e"} Dec 05 09:29:33 crc kubenswrapper[4815]: I1205 09:29:33.175810 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54c4865f77-tppk8" event={"ID":"fceefc10-15e5-4b68-9bb6-e786dde63941","Type":"ContainerStarted","Data":"33c629350d4860d6f2eccfbbcc72d8a43f26d7238112d8661397ea20003c231a"} Dec 05 09:29:33 crc kubenswrapper[4815]: I1205 09:29:33.176183 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54c4865f77-tppk8" event={"ID":"fceefc10-15e5-4b68-9bb6-e786dde63941","Type":"ContainerStarted","Data":"f9ae8ef1f844ccd892a855b7bd92f6ff4da4374c80c419f92cc7583a26c0ba0e"} Dec 05 09:29:33 crc kubenswrapper[4815]: I1205 09:29:33.413446 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 05 09:29:33 crc kubenswrapper[4815]: I1205 09:29:33.474578 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6d38833-a5a9-4961-9ce8-7722b58624a3" path="/var/lib/kubelet/pods/b6d38833-a5a9-4961-9ce8-7722b58624a3/volumes" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.046846 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.140:9311/healthcheck\": read tcp 10.217.0.2:41892->10.217.0.140:9311: read: connection reset by peer" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.047434 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-cd8fb7c76-9g4zb" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.140:9311/healthcheck\": read tcp 10.217.0.2:41896->10.217.0.140:9311: read: connection reset by peer" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.191201 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-54c4865f77-tppk8" event={"ID":"fceefc10-15e5-4b68-9bb6-e786dde63941","Type":"ContainerStarted","Data":"76d630aaabbb3c410085a00480541566477acaa116b5318ab362a10cfac2a29a"} Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.191614 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.193456 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05","Type":"ContainerStarted","Data":"7fa90d2709baeea47f863c5d54b2e29c4448b2729ea50d90504e5710abe91058"} Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.201303 4815 generic.go:334] "Generic (PLEG): container finished" podID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerID="fba393df4f5ab9c99d2166ead3e4b4b3180ea1a121161d1a36e462c3245f3a2a" exitCode=0 Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.201338 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cd8fb7c76-9g4zb" event={"ID":"2acc91d0-1a89-4cef-bcbb-90bb61e1323f","Type":"ContainerDied","Data":"fba393df4f5ab9c99d2166ead3e4b4b3180ea1a121161d1a36e462c3245f3a2a"} Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.226333 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-54c4865f77-tppk8" podStartSLOduration=3.226311221 podStartE2EDuration="3.226311221s" podCreationTimestamp="2025-12-05 09:29:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:34.210366884 +0000 UTC m=+1373.088973741" watchObservedRunningTime="2025-12-05 09:29:34.226311221 +0000 UTC m=+1373.104918048" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.691165 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.835633 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-combined-ca-bundle\") pod \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.835784 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-logs\") pod \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.835827 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data\") pod \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.836231 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c428r\" (UniqueName: \"kubernetes.io/projected/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-kube-api-access-c428r\") pod \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.836262 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data-custom\") pod \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\" (UID: \"2acc91d0-1a89-4cef-bcbb-90bb61e1323f\") " Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.838263 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-logs" (OuterVolumeSpecName: "logs") pod "2acc91d0-1a89-4cef-bcbb-90bb61e1323f" (UID: "2acc91d0-1a89-4cef-bcbb-90bb61e1323f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.842350 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2acc91d0-1a89-4cef-bcbb-90bb61e1323f" (UID: "2acc91d0-1a89-4cef-bcbb-90bb61e1323f"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.850048 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-kube-api-access-c428r" (OuterVolumeSpecName: "kube-api-access-c428r") pod "2acc91d0-1a89-4cef-bcbb-90bb61e1323f" (UID: "2acc91d0-1a89-4cef-bcbb-90bb61e1323f"). InnerVolumeSpecName "kube-api-access-c428r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.874080 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2acc91d0-1a89-4cef-bcbb-90bb61e1323f" (UID: "2acc91d0-1a89-4cef-bcbb-90bb61e1323f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.895399 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data" (OuterVolumeSpecName: "config-data") pod "2acc91d0-1a89-4cef-bcbb-90bb61e1323f" (UID: "2acc91d0-1a89-4cef-bcbb-90bb61e1323f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.939726 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c428r\" (UniqueName: \"kubernetes.io/projected/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-kube-api-access-c428r\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.939756 4815 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.939765 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.939774 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:34 crc kubenswrapper[4815]: I1205 09:29:34.939783 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2acc91d0-1a89-4cef-bcbb-90bb61e1323f-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.216990 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"158b54fd-bab0-4163-8334-2a542378852e","Type":"ContainerStarted","Data":"b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4"} Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.217140 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.222941 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05","Type":"ContainerStarted","Data":"1dc145b807d95c168c5d598c456123d6cf86a0f7b15b5b5d626e5799c8f63786"} Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.228559 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-cd8fb7c76-9g4zb" Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.228435 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-cd8fb7c76-9g4zb" event={"ID":"2acc91d0-1a89-4cef-bcbb-90bb61e1323f","Type":"ContainerDied","Data":"9064091118df2d26038dca148e2c9cd75684bfcd298947a82a2ff47b82c049f4"} Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.232592 4815 scope.go:117] "RemoveContainer" containerID="fba393df4f5ab9c99d2166ead3e4b4b3180ea1a121161d1a36e462c3245f3a2a" Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.245765 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.022269825 podStartE2EDuration="7.245743163s" podCreationTimestamp="2025-12-05 09:29:28 +0000 UTC" firstStartedPulling="2025-12-05 09:29:29.221463494 +0000 UTC m=+1368.100070331" lastFinishedPulling="2025-12-05 09:29:34.444936842 +0000 UTC m=+1373.323543669" observedRunningTime="2025-12-05 09:29:35.244352136 +0000 UTC m=+1374.122958983" watchObservedRunningTime="2025-12-05 09:29:35.245743163 +0000 UTC m=+1374.124350000" Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.306572 4815 scope.go:117] "RemoveContainer" containerID="45ef5ea6df8164df3bad60882166ae470ca705e908389d422fa6fc2185fe4ea0" Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.331325 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-6f56547bd4-2dxhd" Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.341389 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-cd8fb7c76-9g4zb"] Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.352516 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-cd8fb7c76-9g4zb"] Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.439442 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" path="/var/lib/kubelet/pods/2acc91d0-1a89-4cef-bcbb-90bb61e1323f/volumes" Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.573644 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.657092 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85487d4b67-wqr5k"] Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.657353 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" podUID="36611e8e-67fb-420a-a1d6-1ad80e671aa9" containerName="dnsmasq-dns" containerID="cri-o://25adc14737ef33dd8e88043f554f20a7ecad51e478b30ff09764cfc3541be9a2" gracePeriod=10 Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.841650 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.936404 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" podUID="36611e8e-67fb-420a-a1d6-1ad80e671aa9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.139:5353: connect: connection refused" Dec 05 09:29:35 crc kubenswrapper[4815]: I1205 09:29:35.972127 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.277805 4815 generic.go:334] "Generic (PLEG): container finished" podID="36611e8e-67fb-420a-a1d6-1ad80e671aa9" containerID="25adc14737ef33dd8e88043f554f20a7ecad51e478b30ff09764cfc3541be9a2" exitCode=0 Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.277853 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" event={"ID":"36611e8e-67fb-420a-a1d6-1ad80e671aa9","Type":"ContainerDied","Data":"25adc14737ef33dd8e88043f554f20a7ecad51e478b30ff09764cfc3541be9a2"} Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.278071 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" event={"ID":"36611e8e-67fb-420a-a1d6-1ad80e671aa9","Type":"ContainerDied","Data":"5e8297460545b0eeb384e86dad603fb98370ea66556995921e11f01952f98f37"} Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.278087 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e8297460545b0eeb384e86dad603fb98370ea66556995921e11f01952f98f37" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.280523 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05","Type":"ContainerStarted","Data":"9a131c5afd02acce300328904592edbda8dbddb6bc278269980320dc30a35eb2"} Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.281780 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.288778 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" containerName="cinder-scheduler" containerID="cri-o://b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5" gracePeriod=30 Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.288844 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" containerName="probe" containerID="cri-o://b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5" gracePeriod=30 Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.301899 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.301879793 podStartE2EDuration="4.301879793s" podCreationTimestamp="2025-12-05 09:29:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:36.297844052 +0000 UTC m=+1375.176450889" watchObservedRunningTime="2025-12-05 09:29:36.301879793 +0000 UTC m=+1375.180486630" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.309058 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.388353 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-dns-svc\") pod \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.388426 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-sb\") pod \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.388532 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-config\") pod \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.388554 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-nb\") pod \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.388658 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsk25\" (UniqueName: \"kubernetes.io/projected/36611e8e-67fb-420a-a1d6-1ad80e671aa9-kube-api-access-qsk25\") pod \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\" (UID: \"36611e8e-67fb-420a-a1d6-1ad80e671aa9\") " Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.394104 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36611e8e-67fb-420a-a1d6-1ad80e671aa9-kube-api-access-qsk25" (OuterVolumeSpecName: "kube-api-access-qsk25") pod "36611e8e-67fb-420a-a1d6-1ad80e671aa9" (UID: "36611e8e-67fb-420a-a1d6-1ad80e671aa9"). InnerVolumeSpecName "kube-api-access-qsk25". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.467258 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "36611e8e-67fb-420a-a1d6-1ad80e671aa9" (UID: "36611e8e-67fb-420a-a1d6-1ad80e671aa9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.486101 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "36611e8e-67fb-420a-a1d6-1ad80e671aa9" (UID: "36611e8e-67fb-420a-a1d6-1ad80e671aa9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.491144 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsk25\" (UniqueName: \"kubernetes.io/projected/36611e8e-67fb-420a-a1d6-1ad80e671aa9-kube-api-access-qsk25\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.491447 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.491548 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.495269 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-config" (OuterVolumeSpecName: "config") pod "36611e8e-67fb-420a-a1d6-1ad80e671aa9" (UID: "36611e8e-67fb-420a-a1d6-1ad80e671aa9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.520936 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "36611e8e-67fb-420a-a1d6-1ad80e671aa9" (UID: "36611e8e-67fb-420a-a1d6-1ad80e671aa9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.592853 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:36 crc kubenswrapper[4815]: I1205 09:29:36.592887 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/36611e8e-67fb-420a-a1d6-1ad80e671aa9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:37 crc kubenswrapper[4815]: I1205 09:29:37.297414 4815 generic.go:334] "Generic (PLEG): container finished" podID="5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" containerID="b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5" exitCode=0 Dec 05 09:29:37 crc kubenswrapper[4815]: I1205 09:29:37.297626 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891","Type":"ContainerDied","Data":"b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5"} Dec 05 09:29:37 crc kubenswrapper[4815]: I1205 09:29:37.297728 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85487d4b67-wqr5k" Dec 05 09:29:37 crc kubenswrapper[4815]: I1205 09:29:37.326303 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85487d4b67-wqr5k"] Dec 05 09:29:37 crc kubenswrapper[4815]: I1205 09:29:37.337906 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85487d4b67-wqr5k"] Dec 05 09:29:37 crc kubenswrapper[4815]: I1205 09:29:37.429632 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36611e8e-67fb-420a-a1d6-1ad80e671aa9" path="/var/lib/kubelet/pods/36611e8e-67fb-420a-a1d6-1ad80e671aa9/volumes" Dec 05 09:29:37 crc kubenswrapper[4815]: I1205 09:29:37.957080 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.123406 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-etc-machine-id\") pod \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.123754 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x26b7\" (UniqueName: \"kubernetes.io/projected/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-kube-api-access-x26b7\") pod \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.123778 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-scripts\") pod \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.123825 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data-custom\") pod \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.123853 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data\") pod \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.123926 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-combined-ca-bundle\") pod \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\" (UID: \"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891\") " Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.124244 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" (UID: "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.143705 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-kube-api-access-x26b7" (OuterVolumeSpecName: "kube-api-access-x26b7") pod "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" (UID: "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891"). InnerVolumeSpecName "kube-api-access-x26b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.145652 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" (UID: "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.145740 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-scripts" (OuterVolumeSpecName: "scripts") pod "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" (UID: "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.196665 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" (UID: "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.226679 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x26b7\" (UniqueName: \"kubernetes.io/projected/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-kube-api-access-x26b7\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.226709 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.226718 4815 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.226727 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.226735 4815 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.276594 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data" (OuterVolumeSpecName: "config-data") pod "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" (UID: "5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.307137 4815 generic.go:334] "Generic (PLEG): container finished" podID="5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" containerID="b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5" exitCode=0 Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.307202 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891","Type":"ContainerDied","Data":"b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5"} Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.308130 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891","Type":"ContainerDied","Data":"ae2300cddaf4a952d8ee6ec0ff6cd2505a6315dea681093657ffb4ad93db09ce"} Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.308166 4815 scope.go:117] "RemoveContainer" containerID="b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.307222 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.331106 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.353924 4815 scope.go:117] "RemoveContainer" containerID="b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.387480 4815 scope.go:117] "RemoveContainer" containerID="b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.387655 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:29:38 crc kubenswrapper[4815]: E1205 09:29:38.388153 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5\": container with ID starting with b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5 not found: ID does not exist" containerID="b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.388199 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5"} err="failed to get container status \"b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5\": rpc error: code = NotFound desc = could not find container \"b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5\": container with ID starting with b697c9faedac1f3203205d7c18df1c68087dd22317da673688cd18e172107ef5 not found: ID does not exist" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.388233 4815 scope.go:117] "RemoveContainer" containerID="b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5" Dec 05 09:29:38 crc kubenswrapper[4815]: E1205 09:29:38.388674 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5\": container with ID starting with b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5 not found: ID does not exist" containerID="b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.388698 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5"} err="failed to get container status \"b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5\": rpc error: code = NotFound desc = could not find container \"b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5\": container with ID starting with b5da3b30790d06868f6d842a2f8486dd7a8cac4def5d86c1753c2c9cb38a3ad5 not found: ID does not exist" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.435924 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.447308 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:29:38 crc kubenswrapper[4815]: E1205 09:29:38.447900 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api-log" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.447926 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api-log" Dec 05 09:29:38 crc kubenswrapper[4815]: E1205 09:29:38.447937 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36611e8e-67fb-420a-a1d6-1ad80e671aa9" containerName="dnsmasq-dns" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.447946 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="36611e8e-67fb-420a-a1d6-1ad80e671aa9" containerName="dnsmasq-dns" Dec 05 09:29:38 crc kubenswrapper[4815]: E1205 09:29:38.447958 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" containerName="probe" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.447966 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" containerName="probe" Dec 05 09:29:38 crc kubenswrapper[4815]: E1205 09:29:38.447985 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36611e8e-67fb-420a-a1d6-1ad80e671aa9" containerName="init" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.447993 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="36611e8e-67fb-420a-a1d6-1ad80e671aa9" containerName="init" Dec 05 09:29:38 crc kubenswrapper[4815]: E1205 09:29:38.448017 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" containerName="cinder-scheduler" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.448029 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" containerName="cinder-scheduler" Dec 05 09:29:38 crc kubenswrapper[4815]: E1205 09:29:38.448042 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.448050 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.448378 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" containerName="probe" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.448412 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="36611e8e-67fb-420a-a1d6-1ad80e671aa9" containerName="dnsmasq-dns" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.448435 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api-log" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.448448 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" containerName="cinder-scheduler" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.448497 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2acc91d0-1a89-4cef-bcbb-90bb61e1323f" containerName="barbican-api" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.450028 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.452943 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.467259 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.533945 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-config-data\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.533995 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.534021 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-scripts\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.534072 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brk6r\" (UniqueName: \"kubernetes.io/projected/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-kube-api-access-brk6r\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.534126 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.534157 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.635910 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brk6r\" (UniqueName: \"kubernetes.io/projected/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-kube-api-access-brk6r\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.636258 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.636315 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.636368 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-config-data\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.636404 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.636439 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-scripts\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.636592 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.640130 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-scripts\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.640136 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.640862 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-config-data\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.650064 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.662123 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brk6r\" (UniqueName: \"kubernetes.io/projected/9a6d4017-ba13-47e9-b0e2-069a6d6afa21-kube-api-access-brk6r\") pod \"cinder-scheduler-0\" (UID: \"9a6d4017-ba13-47e9-b0e2-069a6d6afa21\") " pod="openstack/cinder-scheduler-0" Dec 05 09:29:38 crc kubenswrapper[4815]: I1205 09:29:38.769332 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.264873 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.318949 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9a6d4017-ba13-47e9-b0e2-069a6d6afa21","Type":"ContainerStarted","Data":"b8f8300dcf0bb36c099d2d2bdd7781fbb3141973751a148f5a81d47f913ca157"} Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.433621 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891" path="/var/lib/kubelet/pods/5a7ca5fc-5e7a-4f7e-a8f1-c9e100fdb891/volumes" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.668573 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.669887 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.673257 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-5gnwf" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.676873 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.677020 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.713798 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.771549 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04cdd6cd-6769-4296-a45d-fffd016fdfd6-combined-ca-bundle\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.771633 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/04cdd6cd-6769-4296-a45d-fffd016fdfd6-openstack-config-secret\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.771682 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2q9p\" (UniqueName: \"kubernetes.io/projected/04cdd6cd-6769-4296-a45d-fffd016fdfd6-kube-api-access-j2q9p\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.771736 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/04cdd6cd-6769-4296-a45d-fffd016fdfd6-openstack-config\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.873946 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04cdd6cd-6769-4296-a45d-fffd016fdfd6-combined-ca-bundle\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.874023 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/04cdd6cd-6769-4296-a45d-fffd016fdfd6-openstack-config-secret\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.874102 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2q9p\" (UniqueName: \"kubernetes.io/projected/04cdd6cd-6769-4296-a45d-fffd016fdfd6-kube-api-access-j2q9p\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.874133 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/04cdd6cd-6769-4296-a45d-fffd016fdfd6-openstack-config\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.874990 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/04cdd6cd-6769-4296-a45d-fffd016fdfd6-openstack-config\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.877229 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/04cdd6cd-6769-4296-a45d-fffd016fdfd6-openstack-config-secret\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.877674 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04cdd6cd-6769-4296-a45d-fffd016fdfd6-combined-ca-bundle\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:39 crc kubenswrapper[4815]: I1205 09:29:39.893390 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2q9p\" (UniqueName: \"kubernetes.io/projected/04cdd6cd-6769-4296-a45d-fffd016fdfd6-kube-api-access-j2q9p\") pod \"openstackclient\" (UID: \"04cdd6cd-6769-4296-a45d-fffd016fdfd6\") " pod="openstack/openstackclient" Dec 05 09:29:40 crc kubenswrapper[4815]: I1205 09:29:40.014336 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 05 09:29:40 crc kubenswrapper[4815]: I1205 09:29:40.364308 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9a6d4017-ba13-47e9-b0e2-069a6d6afa21","Type":"ContainerStarted","Data":"ca40088287da98ddf56df8e47d4c798aeccc2384204c07c6c1daac3739d0709c"} Dec 05 09:29:40 crc kubenswrapper[4815]: I1205 09:29:40.546642 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 05 09:29:41 crc kubenswrapper[4815]: I1205 09:29:41.487718 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"04cdd6cd-6769-4296-a45d-fffd016fdfd6","Type":"ContainerStarted","Data":"4307f03c36075f54a1a85e28e73e70b8baeddd15a57da166991e35850125c8ab"} Dec 05 09:29:41 crc kubenswrapper[4815]: I1205 09:29:41.487946 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9a6d4017-ba13-47e9-b0e2-069a6d6afa21","Type":"ContainerStarted","Data":"43d74ed13af42604b948dfc1a7e816f93f7e76919a8309405737842cee19577f"} Dec 05 09:29:41 crc kubenswrapper[4815]: I1205 09:29:41.566396 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.566376946 podStartE2EDuration="3.566376946s" podCreationTimestamp="2025-12-05 09:29:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:41.563565958 +0000 UTC m=+1380.442172805" watchObservedRunningTime="2025-12-05 09:29:41.566376946 +0000 UTC m=+1380.444983783" Dec 05 09:29:43 crc kubenswrapper[4815]: I1205 09:29:43.770202 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 05 09:29:46 crc kubenswrapper[4815]: I1205 09:29:46.560962 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 05 09:29:49 crc kubenswrapper[4815]: I1205 09:29:49.159682 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 05 09:29:50 crc kubenswrapper[4815]: I1205 09:29:50.191875 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:29:50 crc kubenswrapper[4815]: I1205 09:29:50.191927 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:29:54 crc kubenswrapper[4815]: I1205 09:29:54.638166 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:29:55 crc kubenswrapper[4815]: I1205 09:29:55.089656 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:29:55 crc kubenswrapper[4815]: I1205 09:29:55.090016 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="ceilometer-central-agent" containerID="cri-o://fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3" gracePeriod=30 Dec 05 09:29:55 crc kubenswrapper[4815]: I1205 09:29:55.090085 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="sg-core" containerID="cri-o://13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e" gracePeriod=30 Dec 05 09:29:55 crc kubenswrapper[4815]: I1205 09:29:55.090161 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="ceilometer-notification-agent" containerID="cri-o://66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6" gracePeriod=30 Dec 05 09:29:55 crc kubenswrapper[4815]: I1205 09:29:55.090165 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="proxy-httpd" containerID="cri-o://b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4" gracePeriod=30 Dec 05 09:29:55 crc kubenswrapper[4815]: I1205 09:29:55.108804 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.148:3000/\": EOF" Dec 05 09:29:55 crc kubenswrapper[4815]: I1205 09:29:55.628293 4815 generic.go:334] "Generic (PLEG): container finished" podID="158b54fd-bab0-4163-8334-2a542378852e" containerID="b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4" exitCode=0 Dec 05 09:29:55 crc kubenswrapper[4815]: I1205 09:29:55.628685 4815 generic.go:334] "Generic (PLEG): container finished" podID="158b54fd-bab0-4163-8334-2a542378852e" containerID="13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e" exitCode=2 Dec 05 09:29:55 crc kubenswrapper[4815]: I1205 09:29:55.628570 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"158b54fd-bab0-4163-8334-2a542378852e","Type":"ContainerDied","Data":"b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4"} Dec 05 09:29:55 crc kubenswrapper[4815]: I1205 09:29:55.628742 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"158b54fd-bab0-4163-8334-2a542378852e","Type":"ContainerDied","Data":"13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e"} Dec 05 09:29:56 crc kubenswrapper[4815]: I1205 09:29:56.639698 4815 generic.go:334] "Generic (PLEG): container finished" podID="158b54fd-bab0-4163-8334-2a542378852e" containerID="fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3" exitCode=0 Dec 05 09:29:56 crc kubenswrapper[4815]: I1205 09:29:56.639778 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"158b54fd-bab0-4163-8334-2a542378852e","Type":"ContainerDied","Data":"fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3"} Dec 05 09:29:56 crc kubenswrapper[4815]: I1205 09:29:56.643177 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"04cdd6cd-6769-4296-a45d-fffd016fdfd6","Type":"ContainerStarted","Data":"a29624f61cb59f05feadfe536ba6c1ed00f17dbfc3643c12f809c519724bcef6"} Dec 05 09:29:56 crc kubenswrapper[4815]: I1205 09:29:56.673740 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.848875189 podStartE2EDuration="17.673715631s" podCreationTimestamp="2025-12-05 09:29:39 +0000 UTC" firstStartedPulling="2025-12-05 09:29:40.566173321 +0000 UTC m=+1379.444780158" lastFinishedPulling="2025-12-05 09:29:55.391013763 +0000 UTC m=+1394.269620600" observedRunningTime="2025-12-05 09:29:56.661330002 +0000 UTC m=+1395.539936839" watchObservedRunningTime="2025-12-05 09:29:56.673715631 +0000 UTC m=+1395.552322468" Dec 05 09:29:56 crc kubenswrapper[4815]: I1205 09:29:56.771240 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.288228 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-qgqvs"] Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.289766 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qgqvs" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.306661 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-qgqvs"] Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.328769 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6b4wq\" (UniqueName: \"kubernetes.io/projected/be43b898-0145-4ee3-aeb0-0e35acaa094d-kube-api-access-6b4wq\") pod \"nova-api-db-create-qgqvs\" (UID: \"be43b898-0145-4ee3-aeb0-0e35acaa094d\") " pod="openstack/nova-api-db-create-qgqvs" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.329047 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be43b898-0145-4ee3-aeb0-0e35acaa094d-operator-scripts\") pod \"nova-api-db-create-qgqvs\" (UID: \"be43b898-0145-4ee3-aeb0-0e35acaa094d\") " pod="openstack/nova-api-db-create-qgqvs" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.431583 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6b4wq\" (UniqueName: \"kubernetes.io/projected/be43b898-0145-4ee3-aeb0-0e35acaa094d-kube-api-access-6b4wq\") pod \"nova-api-db-create-qgqvs\" (UID: \"be43b898-0145-4ee3-aeb0-0e35acaa094d\") " pod="openstack/nova-api-db-create-qgqvs" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.431667 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be43b898-0145-4ee3-aeb0-0e35acaa094d-operator-scripts\") pod \"nova-api-db-create-qgqvs\" (UID: \"be43b898-0145-4ee3-aeb0-0e35acaa094d\") " pod="openstack/nova-api-db-create-qgqvs" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.432414 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be43b898-0145-4ee3-aeb0-0e35acaa094d-operator-scripts\") pod \"nova-api-db-create-qgqvs\" (UID: \"be43b898-0145-4ee3-aeb0-0e35acaa094d\") " pod="openstack/nova-api-db-create-qgqvs" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.456445 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-88f59"] Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.460513 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6b4wq\" (UniqueName: \"kubernetes.io/projected/be43b898-0145-4ee3-aeb0-0e35acaa094d-kube-api-access-6b4wq\") pod \"nova-api-db-create-qgqvs\" (UID: \"be43b898-0145-4ee3-aeb0-0e35acaa094d\") " pod="openstack/nova-api-db-create-qgqvs" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.468883 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-88f59" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.483528 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-88f59"] Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.534598 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0186806-095c-4433-bb68-55732c995bed-operator-scripts\") pod \"nova-cell0-db-create-88f59\" (UID: \"e0186806-095c-4433-bb68-55732c995bed\") " pod="openstack/nova-cell0-db-create-88f59" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.534651 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8rtr\" (UniqueName: \"kubernetes.io/projected/e0186806-095c-4433-bb68-55732c995bed-kube-api-access-l8rtr\") pod \"nova-cell0-db-create-88f59\" (UID: \"e0186806-095c-4433-bb68-55732c995bed\") " pod="openstack/nova-cell0-db-create-88f59" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.543048 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-41a2-account-create-update-rxfbj"] Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.554811 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-41a2-account-create-update-rxfbj" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.559594 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.576987 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-41a2-account-create-update-rxfbj"] Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.628866 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qgqvs" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.636541 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-operator-scripts\") pod \"nova-api-41a2-account-create-update-rxfbj\" (UID: \"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c\") " pod="openstack/nova-api-41a2-account-create-update-rxfbj" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.636638 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0186806-095c-4433-bb68-55732c995bed-operator-scripts\") pod \"nova-cell0-db-create-88f59\" (UID: \"e0186806-095c-4433-bb68-55732c995bed\") " pod="openstack/nova-cell0-db-create-88f59" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.636670 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8rtr\" (UniqueName: \"kubernetes.io/projected/e0186806-095c-4433-bb68-55732c995bed-kube-api-access-l8rtr\") pod \"nova-cell0-db-create-88f59\" (UID: \"e0186806-095c-4433-bb68-55732c995bed\") " pod="openstack/nova-cell0-db-create-88f59" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.636700 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph4k4\" (UniqueName: \"kubernetes.io/projected/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-kube-api-access-ph4k4\") pod \"nova-api-41a2-account-create-update-rxfbj\" (UID: \"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c\") " pod="openstack/nova-api-41a2-account-create-update-rxfbj" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.637472 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0186806-095c-4433-bb68-55732c995bed-operator-scripts\") pod \"nova-cell0-db-create-88f59\" (UID: \"e0186806-095c-4433-bb68-55732c995bed\") " pod="openstack/nova-cell0-db-create-88f59" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.643937 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-jmvvv"] Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.645477 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jmvvv" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.690530 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jmvvv"] Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.720156 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8rtr\" (UniqueName: \"kubernetes.io/projected/e0186806-095c-4433-bb68-55732c995bed-kube-api-access-l8rtr\") pod \"nova-cell0-db-create-88f59\" (UID: \"e0186806-095c-4433-bb68-55732c995bed\") " pod="openstack/nova-cell0-db-create-88f59" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.742466 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5de674b4-317a-4d40-959c-6e0b050a18de-operator-scripts\") pod \"nova-cell1-db-create-jmvvv\" (UID: \"5de674b4-317a-4d40-959c-6e0b050a18de\") " pod="openstack/nova-cell1-db-create-jmvvv" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.742596 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-operator-scripts\") pod \"nova-api-41a2-account-create-update-rxfbj\" (UID: \"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c\") " pod="openstack/nova-api-41a2-account-create-update-rxfbj" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.742665 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6g92w\" (UniqueName: \"kubernetes.io/projected/5de674b4-317a-4d40-959c-6e0b050a18de-kube-api-access-6g92w\") pod \"nova-cell1-db-create-jmvvv\" (UID: \"5de674b4-317a-4d40-959c-6e0b050a18de\") " pod="openstack/nova-cell1-db-create-jmvvv" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.742737 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph4k4\" (UniqueName: \"kubernetes.io/projected/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-kube-api-access-ph4k4\") pod \"nova-api-41a2-account-create-update-rxfbj\" (UID: \"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c\") " pod="openstack/nova-api-41a2-account-create-update-rxfbj" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.743468 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-operator-scripts\") pod \"nova-api-41a2-account-create-update-rxfbj\" (UID: \"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c\") " pod="openstack/nova-api-41a2-account-create-update-rxfbj" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.785221 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph4k4\" (UniqueName: \"kubernetes.io/projected/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-kube-api-access-ph4k4\") pod \"nova-api-41a2-account-create-update-rxfbj\" (UID: \"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c\") " pod="openstack/nova-api-41a2-account-create-update-rxfbj" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.844025 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5de674b4-317a-4d40-959c-6e0b050a18de-operator-scripts\") pod \"nova-cell1-db-create-jmvvv\" (UID: \"5de674b4-317a-4d40-959c-6e0b050a18de\") " pod="openstack/nova-cell1-db-create-jmvvv" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.844402 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6g92w\" (UniqueName: \"kubernetes.io/projected/5de674b4-317a-4d40-959c-6e0b050a18de-kube-api-access-6g92w\") pod \"nova-cell1-db-create-jmvvv\" (UID: \"5de674b4-317a-4d40-959c-6e0b050a18de\") " pod="openstack/nova-cell1-db-create-jmvvv" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.844885 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5de674b4-317a-4d40-959c-6e0b050a18de-operator-scripts\") pod \"nova-cell1-db-create-jmvvv\" (UID: \"5de674b4-317a-4d40-959c-6e0b050a18de\") " pod="openstack/nova-cell1-db-create-jmvvv" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.845053 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-88f59" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.868741 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-d1d3-account-create-update-5gxdr"] Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.869840 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.872532 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.876865 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-41a2-account-create-update-rxfbj" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.887205 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-d1d3-account-create-update-5gxdr"] Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.894649 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6g92w\" (UniqueName: \"kubernetes.io/projected/5de674b4-317a-4d40-959c-6e0b050a18de-kube-api-access-6g92w\") pod \"nova-cell1-db-create-jmvvv\" (UID: \"5de674b4-317a-4d40-959c-6e0b050a18de\") " pod="openstack/nova-cell1-db-create-jmvvv" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.947995 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnxvh\" (UniqueName: \"kubernetes.io/projected/15e7e2fa-915a-407e-a8cd-1a0225692722-kube-api-access-bnxvh\") pod \"nova-cell0-d1d3-account-create-update-5gxdr\" (UID: \"15e7e2fa-915a-407e-a8cd-1a0225692722\") " pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" Dec 05 09:29:57 crc kubenswrapper[4815]: I1205 09:29:57.948072 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15e7e2fa-915a-407e-a8cd-1a0225692722-operator-scripts\") pod \"nova-cell0-d1d3-account-create-update-5gxdr\" (UID: \"15e7e2fa-915a-407e-a8cd-1a0225692722\") " pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.050285 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnxvh\" (UniqueName: \"kubernetes.io/projected/15e7e2fa-915a-407e-a8cd-1a0225692722-kube-api-access-bnxvh\") pod \"nova-cell0-d1d3-account-create-update-5gxdr\" (UID: \"15e7e2fa-915a-407e-a8cd-1a0225692722\") " pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.050579 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15e7e2fa-915a-407e-a8cd-1a0225692722-operator-scripts\") pod \"nova-cell0-d1d3-account-create-update-5gxdr\" (UID: \"15e7e2fa-915a-407e-a8cd-1a0225692722\") " pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.052317 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15e7e2fa-915a-407e-a8cd-1a0225692722-operator-scripts\") pod \"nova-cell0-d1d3-account-create-update-5gxdr\" (UID: \"15e7e2fa-915a-407e-a8cd-1a0225692722\") " pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.094107 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-2bb0-account-create-update-f4zbp"] Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.118812 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnxvh\" (UniqueName: \"kubernetes.io/projected/15e7e2fa-915a-407e-a8cd-1a0225692722-kube-api-access-bnxvh\") pod \"nova-cell0-d1d3-account-create-update-5gxdr\" (UID: \"15e7e2fa-915a-407e-a8cd-1a0225692722\") " pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.120512 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.131377 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.144369 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jmvvv" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.162132 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d48f238-7fdf-4593-92e3-352f07e26062-operator-scripts\") pod \"nova-cell1-2bb0-account-create-update-f4zbp\" (UID: \"2d48f238-7fdf-4593-92e3-352f07e26062\") " pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.162178 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdx7s\" (UniqueName: \"kubernetes.io/projected/2d48f238-7fdf-4593-92e3-352f07e26062-kube-api-access-qdx7s\") pod \"nova-cell1-2bb0-account-create-update-f4zbp\" (UID: \"2d48f238-7fdf-4593-92e3-352f07e26062\") " pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.174292 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-549777fc4b-mjkbk" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.208322 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2bb0-account-create-update-f4zbp"] Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.249352 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.267506 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d48f238-7fdf-4593-92e3-352f07e26062-operator-scripts\") pod \"nova-cell1-2bb0-account-create-update-f4zbp\" (UID: \"2d48f238-7fdf-4593-92e3-352f07e26062\") " pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.267567 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdx7s\" (UniqueName: \"kubernetes.io/projected/2d48f238-7fdf-4593-92e3-352f07e26062-kube-api-access-qdx7s\") pod \"nova-cell1-2bb0-account-create-update-f4zbp\" (UID: \"2d48f238-7fdf-4593-92e3-352f07e26062\") " pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.269378 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d48f238-7fdf-4593-92e3-352f07e26062-operator-scripts\") pod \"nova-cell1-2bb0-account-create-update-f4zbp\" (UID: \"2d48f238-7fdf-4593-92e3-352f07e26062\") " pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.304137 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdx7s\" (UniqueName: \"kubernetes.io/projected/2d48f238-7fdf-4593-92e3-352f07e26062-kube-api-access-qdx7s\") pod \"nova-cell1-2bb0-account-create-update-f4zbp\" (UID: \"2d48f238-7fdf-4593-92e3-352f07e26062\") " pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.472928 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.559643 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.148:3000/\": dial tcp 10.217.0.148:3000: connect: connection refused" Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.735189 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-qgqvs"] Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.930593 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-88f59"] Dec 05 09:29:58 crc kubenswrapper[4815]: W1205 09:29:58.934761 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0186806_095c_4433_bb68_55732c995bed.slice/crio-7191108af0eb96c1520a648d1824d816dc9fd40d1f3e779770a805164faa1060 WatchSource:0}: Error finding container 7191108af0eb96c1520a648d1824d816dc9fd40d1f3e779770a805164faa1060: Status 404 returned error can't find the container with id 7191108af0eb96c1520a648d1824d816dc9fd40d1f3e779770a805164faa1060 Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.966037 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-41a2-account-create-update-rxfbj"] Dec 05 09:29:58 crc kubenswrapper[4815]: I1205 09:29:58.991856 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-jmvvv"] Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.284048 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-d1d3-account-create-update-5gxdr"] Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.474772 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2bb0-account-create-update-f4zbp"] Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.669871 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.729316 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-sg-core-conf-yaml\") pod \"158b54fd-bab0-4163-8334-2a542378852e\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.729670 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6f7m\" (UniqueName: \"kubernetes.io/projected/158b54fd-bab0-4163-8334-2a542378852e-kube-api-access-p6f7m\") pod \"158b54fd-bab0-4163-8334-2a542378852e\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.729702 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-run-httpd\") pod \"158b54fd-bab0-4163-8334-2a542378852e\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.729782 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-log-httpd\") pod \"158b54fd-bab0-4163-8334-2a542378852e\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.729811 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-combined-ca-bundle\") pod \"158b54fd-bab0-4163-8334-2a542378852e\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.729880 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-config-data\") pod \"158b54fd-bab0-4163-8334-2a542378852e\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.729947 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-scripts\") pod \"158b54fd-bab0-4163-8334-2a542378852e\" (UID: \"158b54fd-bab0-4163-8334-2a542378852e\") " Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.730540 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "158b54fd-bab0-4163-8334-2a542378852e" (UID: "158b54fd-bab0-4163-8334-2a542378852e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.730736 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "158b54fd-bab0-4163-8334-2a542378852e" (UID: "158b54fd-bab0-4163-8334-2a542378852e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.736928 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-88f59" event={"ID":"e0186806-095c-4433-bb68-55732c995bed","Type":"ContainerStarted","Data":"5959f1dc92563e332c1be778e899ecdc1f1000da1aec244ee4903d2aad1ce66d"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.736987 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-88f59" event={"ID":"e0186806-095c-4433-bb68-55732c995bed","Type":"ContainerStarted","Data":"7191108af0eb96c1520a648d1824d816dc9fd40d1f3e779770a805164faa1060"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.752556 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" event={"ID":"2d48f238-7fdf-4593-92e3-352f07e26062","Type":"ContainerStarted","Data":"70efe71fc22921e9ee5167ae6b525f493c32601e76a9936b1d23b7f03f3a815b"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.753179 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/158b54fd-bab0-4163-8334-2a542378852e-kube-api-access-p6f7m" (OuterVolumeSpecName: "kube-api-access-p6f7m") pod "158b54fd-bab0-4163-8334-2a542378852e" (UID: "158b54fd-bab0-4163-8334-2a542378852e"). InnerVolumeSpecName "kube-api-access-p6f7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.769927 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-88f59" podStartSLOduration=2.769907137 podStartE2EDuration="2.769907137s" podCreationTimestamp="2025-12-05 09:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:59.763107131 +0000 UTC m=+1398.641713968" watchObservedRunningTime="2025-12-05 09:29:59.769907137 +0000 UTC m=+1398.648513974" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.778941 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" event={"ID":"15e7e2fa-915a-407e-a8cd-1a0225692722","Type":"ContainerStarted","Data":"6c4d79fa7328539efe6954bcf59e9b5f6248c0007449f08a17fddd519d0c0169"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.779612 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-scripts" (OuterVolumeSpecName: "scripts") pod "158b54fd-bab0-4163-8334-2a542378852e" (UID: "158b54fd-bab0-4163-8334-2a542378852e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.826279 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jmvvv" event={"ID":"5de674b4-317a-4d40-959c-6e0b050a18de","Type":"ContainerStarted","Data":"158b471d9b8af04deba859b22bffd924e33ea5f08b023a3d7328196857b8eacf"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.826486 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jmvvv" event={"ID":"5de674b4-317a-4d40-959c-6e0b050a18de","Type":"ContainerStarted","Data":"41d3a8ebd821df59eb084c4bbb407faea8cfd222cc9aaa1e6768cc5beb59523a"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.832798 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6f7m\" (UniqueName: \"kubernetes.io/projected/158b54fd-bab0-4163-8334-2a542378852e-kube-api-access-p6f7m\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.832825 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.832834 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/158b54fd-bab0-4163-8334-2a542378852e-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.832842 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.844230 4815 generic.go:334] "Generic (PLEG): container finished" podID="158b54fd-bab0-4163-8334-2a542378852e" containerID="66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6" exitCode=0 Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.844312 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"158b54fd-bab0-4163-8334-2a542378852e","Type":"ContainerDied","Data":"66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.844341 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"158b54fd-bab0-4163-8334-2a542378852e","Type":"ContainerDied","Data":"fd11f7591d0ef77e77394b1e130b9dcb7c89d1264ab02d5a13b6b5628ac52bca"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.844362 4815 scope.go:117] "RemoveContainer" containerID="b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.847861 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.853194 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-jmvvv" podStartSLOduration=2.8531793629999997 podStartE2EDuration="2.853179363s" podCreationTimestamp="2025-12-05 09:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:59.851846037 +0000 UTC m=+1398.730452864" watchObservedRunningTime="2025-12-05 09:29:59.853179363 +0000 UTC m=+1398.731786200" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.857718 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-qgqvs" event={"ID":"be43b898-0145-4ee3-aeb0-0e35acaa094d","Type":"ContainerStarted","Data":"46f2696aa70c8ba1f1e30cd56782981a9f0fbdffd5d57038b0d543109e345293"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.857774 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-qgqvs" event={"ID":"be43b898-0145-4ee3-aeb0-0e35acaa094d","Type":"ContainerStarted","Data":"86322404e62f0b6ad6046cd3bfe83b1b961519b8bd0c9f90fce852b9a46ffbd4"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.892046 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-41a2-account-create-update-rxfbj" event={"ID":"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c","Type":"ContainerStarted","Data":"71d7348eefb9baa02cf9473eb142136a4611deeb4abd2da64031b8f3fc0c03fe"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.892096 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-41a2-account-create-update-rxfbj" event={"ID":"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c","Type":"ContainerStarted","Data":"d99327d129745ab5888a16c30546eef5e75079e33f31f724ca719314b8c03678"} Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.940211 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-qgqvs" podStartSLOduration=2.940192131 podStartE2EDuration="2.940192131s" podCreationTimestamp="2025-12-05 09:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:59.887478044 +0000 UTC m=+1398.766084881" watchObservedRunningTime="2025-12-05 09:29:59.940192131 +0000 UTC m=+1398.818798968" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.953140 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "158b54fd-bab0-4163-8334-2a542378852e" (UID: "158b54fd-bab0-4163-8334-2a542378852e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:29:59 crc kubenswrapper[4815]: I1205 09:29:59.960004 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-41a2-account-create-update-rxfbj" podStartSLOduration=2.959982944 podStartE2EDuration="2.959982944s" podCreationTimestamp="2025-12-05 09:29:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:29:59.925278332 +0000 UTC m=+1398.803885169" watchObservedRunningTime="2025-12-05 09:29:59.959982944 +0000 UTC m=+1398.838589781" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.043209 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.056121 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "158b54fd-bab0-4163-8334-2a542378852e" (UID: "158b54fd-bab0-4163-8334-2a542378852e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.068447 4815 scope.go:117] "RemoveContainer" containerID="13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.091212 4815 scope.go:117] "RemoveContainer" containerID="66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.104228 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-config-data" (OuterVolumeSpecName: "config-data") pod "158b54fd-bab0-4163-8334-2a542378852e" (UID: "158b54fd-bab0-4163-8334-2a542378852e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.144562 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.144603 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/158b54fd-bab0-4163-8334-2a542378852e-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.145450 4815 scope.go:117] "RemoveContainer" containerID="fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.154662 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8"] Dec 05 09:30:00 crc kubenswrapper[4815]: E1205 09:30:00.155050 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="ceilometer-notification-agent" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.155063 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="ceilometer-notification-agent" Dec 05 09:30:00 crc kubenswrapper[4815]: E1205 09:30:00.155074 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="proxy-httpd" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.155081 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="proxy-httpd" Dec 05 09:30:00 crc kubenswrapper[4815]: E1205 09:30:00.155109 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="ceilometer-central-agent" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.155115 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="ceilometer-central-agent" Dec 05 09:30:00 crc kubenswrapper[4815]: E1205 09:30:00.155131 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="sg-core" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.155137 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="sg-core" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.155319 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="sg-core" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.155371 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="proxy-httpd" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.155384 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="ceilometer-notification-agent" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.155391 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="158b54fd-bab0-4163-8334-2a542378852e" containerName="ceilometer-central-agent" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.194148 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.195422 4815 scope.go:117] "RemoveContainer" containerID="b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4" Dec 05 09:30:00 crc kubenswrapper[4815]: E1205 09:30:00.196671 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4\": container with ID starting with b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4 not found: ID does not exist" containerID="b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.196711 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4"} err="failed to get container status \"b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4\": rpc error: code = NotFound desc = could not find container \"b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4\": container with ID starting with b804272f1ced339888a08463ed4d872b0b892738f765db8d2019dd802c2e9fc4 not found: ID does not exist" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.196746 4815 scope.go:117] "RemoveContainer" containerID="13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e" Dec 05 09:30:00 crc kubenswrapper[4815]: E1205 09:30:00.197289 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e\": container with ID starting with 13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e not found: ID does not exist" containerID="13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.197335 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e"} err="failed to get container status \"13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e\": rpc error: code = NotFound desc = could not find container \"13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e\": container with ID starting with 13114766b519e5c4a6af78deae4d76bdf88fb12b7e7b94dd6f51b10f01654a3e not found: ID does not exist" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.197366 4815 scope.go:117] "RemoveContainer" containerID="66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.197754 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 09:30:00 crc kubenswrapper[4815]: E1205 09:30:00.199226 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6\": container with ID starting with 66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6 not found: ID does not exist" containerID="66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.199288 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6"} err="failed to get container status \"66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6\": rpc error: code = NotFound desc = could not find container \"66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6\": container with ID starting with 66ccb5aa089c619b92e75d02216b6e57c4192178e77af2feb6dd49b4958370b6 not found: ID does not exist" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.199310 4815 scope.go:117] "RemoveContainer" containerID="fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3" Dec 05 09:30:00 crc kubenswrapper[4815]: E1205 09:30:00.199619 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3\": container with ID starting with fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3 not found: ID does not exist" containerID="fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.199649 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3"} err="failed to get container status \"fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3\": rpc error: code = NotFound desc = could not find container \"fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3\": container with ID starting with fe7b8b4155f5609fa25c0fcf09409b8fbd90eb1763d0056a6bbd52bdb2fedda3 not found: ID does not exist" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.205181 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.209745 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8"] Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.256823 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.266408 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.281234 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.292151 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.294960 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.295326 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.295410 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.348611 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-config-data\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.348675 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b2189a3-9b89-4167-91ce-b9e92a360d11-secret-volume\") pod \"collect-profiles-29415450-mxlg8\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.348699 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6p5z\" (UniqueName: \"kubernetes.io/projected/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-kube-api-access-p6p5z\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.348719 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-scripts\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.348786 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.348822 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-log-httpd\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.348847 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwbhs\" (UniqueName: \"kubernetes.io/projected/0b2189a3-9b89-4167-91ce-b9e92a360d11-kube-api-access-kwbhs\") pod \"collect-profiles-29415450-mxlg8\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.348939 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-run-httpd\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.348977 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b2189a3-9b89-4167-91ce-b9e92a360d11-config-volume\") pod \"collect-profiles-29415450-mxlg8\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.349011 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.449907 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.450271 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-config-data\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.450314 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b2189a3-9b89-4167-91ce-b9e92a360d11-secret-volume\") pod \"collect-profiles-29415450-mxlg8\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.450338 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6p5z\" (UniqueName: \"kubernetes.io/projected/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-kube-api-access-p6p5z\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.450360 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-scripts\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.450430 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.450463 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-log-httpd\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.450506 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwbhs\" (UniqueName: \"kubernetes.io/projected/0b2189a3-9b89-4167-91ce-b9e92a360d11-kube-api-access-kwbhs\") pod \"collect-profiles-29415450-mxlg8\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.450551 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-run-httpd\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.450594 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b2189a3-9b89-4167-91ce-b9e92a360d11-config-volume\") pod \"collect-profiles-29415450-mxlg8\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.450999 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-log-httpd\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.451474 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-run-httpd\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.451559 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b2189a3-9b89-4167-91ce-b9e92a360d11-config-volume\") pod \"collect-profiles-29415450-mxlg8\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.454937 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b2189a3-9b89-4167-91ce-b9e92a360d11-secret-volume\") pod \"collect-profiles-29415450-mxlg8\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.455983 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.456588 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.457189 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-scripts\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.461925 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-config-data\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.474385 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwbhs\" (UniqueName: \"kubernetes.io/projected/0b2189a3-9b89-4167-91ce-b9e92a360d11-kube-api-access-kwbhs\") pod \"collect-profiles-29415450-mxlg8\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.476509 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6p5z\" (UniqueName: \"kubernetes.io/projected/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-kube-api-access-p6p5z\") pod \"ceilometer-0\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.515598 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.608819 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.920721 4815 generic.go:334] "Generic (PLEG): container finished" podID="e0186806-095c-4433-bb68-55732c995bed" containerID="5959f1dc92563e332c1be778e899ecdc1f1000da1aec244ee4903d2aad1ce66d" exitCode=0 Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.921119 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-88f59" event={"ID":"e0186806-095c-4433-bb68-55732c995bed","Type":"ContainerDied","Data":"5959f1dc92563e332c1be778e899ecdc1f1000da1aec244ee4903d2aad1ce66d"} Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.923301 4815 generic.go:334] "Generic (PLEG): container finished" podID="2d48f238-7fdf-4593-92e3-352f07e26062" containerID="e7948e689586307da965d0d523ff5693636db131adf753f832413cf37fcde6d8" exitCode=0 Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.923343 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" event={"ID":"2d48f238-7fdf-4593-92e3-352f07e26062","Type":"ContainerDied","Data":"e7948e689586307da965d0d523ff5693636db131adf753f832413cf37fcde6d8"} Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.938764 4815 generic.go:334] "Generic (PLEG): container finished" podID="15e7e2fa-915a-407e-a8cd-1a0225692722" containerID="61824a91f472cb2f73bb1877b3635263a8793a54054fe0745a6ea6af0604561a" exitCode=0 Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.938865 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" event={"ID":"15e7e2fa-915a-407e-a8cd-1a0225692722","Type":"ContainerDied","Data":"61824a91f472cb2f73bb1877b3635263a8793a54054fe0745a6ea6af0604561a"} Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.963796 4815 generic.go:334] "Generic (PLEG): container finished" podID="5de674b4-317a-4d40-959c-6e0b050a18de" containerID="158b471d9b8af04deba859b22bffd924e33ea5f08b023a3d7328196857b8eacf" exitCode=0 Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.963931 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jmvvv" event={"ID":"5de674b4-317a-4d40-959c-6e0b050a18de","Type":"ContainerDied","Data":"158b471d9b8af04deba859b22bffd924e33ea5f08b023a3d7328196857b8eacf"} Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.981279 4815 generic.go:334] "Generic (PLEG): container finished" podID="be43b898-0145-4ee3-aeb0-0e35acaa094d" containerID="46f2696aa70c8ba1f1e30cd56782981a9f0fbdffd5d57038b0d543109e345293" exitCode=0 Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.981349 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-qgqvs" event={"ID":"be43b898-0145-4ee3-aeb0-0e35acaa094d","Type":"ContainerDied","Data":"46f2696aa70c8ba1f1e30cd56782981a9f0fbdffd5d57038b0d543109e345293"} Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.982959 4815 generic.go:334] "Generic (PLEG): container finished" podID="b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c" containerID="71d7348eefb9baa02cf9473eb142136a4611deeb4abd2da64031b8f3fc0c03fe" exitCode=0 Dec 05 09:30:00 crc kubenswrapper[4815]: I1205 09:30:00.982988 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-41a2-account-create-update-rxfbj" event={"ID":"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c","Type":"ContainerDied","Data":"71d7348eefb9baa02cf9473eb142136a4611deeb4abd2da64031b8f3fc0c03fe"} Dec 05 09:30:01 crc kubenswrapper[4815]: I1205 09:30:01.081211 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8"] Dec 05 09:30:01 crc kubenswrapper[4815]: I1205 09:30:01.409109 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-54c4865f77-tppk8" Dec 05 09:30:01 crc kubenswrapper[4815]: I1205 09:30:01.430747 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="158b54fd-bab0-4163-8334-2a542378852e" path="/var/lib/kubelet/pods/158b54fd-bab0-4163-8334-2a542378852e/volumes" Dec 05 09:30:01 crc kubenswrapper[4815]: I1205 09:30:01.517899 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6bdb56b95d-k5pvr"] Dec 05 09:30:01 crc kubenswrapper[4815]: I1205 09:30:01.518097 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6bdb56b95d-k5pvr" podUID="7126486e-29e9-487c-b603-7bf23251591b" containerName="neutron-api" containerID="cri-o://697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab" gracePeriod=30 Dec 05 09:30:01 crc kubenswrapper[4815]: I1205 09:30:01.518200 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6bdb56b95d-k5pvr" podUID="7126486e-29e9-487c-b603-7bf23251591b" containerName="neutron-httpd" containerID="cri-o://9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35" gracePeriod=30 Dec 05 09:30:01 crc kubenswrapper[4815]: I1205 09:30:01.616811 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:01 crc kubenswrapper[4815]: I1205 09:30:01.995706 4815 generic.go:334] "Generic (PLEG): container finished" podID="7126486e-29e9-487c-b603-7bf23251591b" containerID="9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35" exitCode=0 Dec 05 09:30:01 crc kubenswrapper[4815]: I1205 09:30:01.995830 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdb56b95d-k5pvr" event={"ID":"7126486e-29e9-487c-b603-7bf23251591b","Type":"ContainerDied","Data":"9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35"} Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:01.999699 4815 generic.go:334] "Generic (PLEG): container finished" podID="0b2189a3-9b89-4167-91ce-b9e92a360d11" containerID="ff2675db2ab534dc3b503cf7eed5658aae5dbe001a549df516f0c67935bc53dd" exitCode=0 Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.000311 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" event={"ID":"0b2189a3-9b89-4167-91ce-b9e92a360d11","Type":"ContainerDied","Data":"ff2675db2ab534dc3b503cf7eed5658aae5dbe001a549df516f0c67935bc53dd"} Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.000410 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" event={"ID":"0b2189a3-9b89-4167-91ce-b9e92a360d11","Type":"ContainerStarted","Data":"57051021e529691c600359680419276c5b6b766567beef2898b57a9dd9745b24"} Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.001755 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e03913c-e3a4-42e1-b133-5ef5f95d7b10","Type":"ContainerStarted","Data":"0d93a1fc85ac4130a6fa340c246b3caf51a93e2f746cb15228dc90c7f94e2238"} Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.337141 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.434920 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdx7s\" (UniqueName: \"kubernetes.io/projected/2d48f238-7fdf-4593-92e3-352f07e26062-kube-api-access-qdx7s\") pod \"2d48f238-7fdf-4593-92e3-352f07e26062\" (UID: \"2d48f238-7fdf-4593-92e3-352f07e26062\") " Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.435356 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d48f238-7fdf-4593-92e3-352f07e26062-operator-scripts\") pod \"2d48f238-7fdf-4593-92e3-352f07e26062\" (UID: \"2d48f238-7fdf-4593-92e3-352f07e26062\") " Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.437189 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d48f238-7fdf-4593-92e3-352f07e26062-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d48f238-7fdf-4593-92e3-352f07e26062" (UID: "2d48f238-7fdf-4593-92e3-352f07e26062"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.452997 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d48f238-7fdf-4593-92e3-352f07e26062-kube-api-access-qdx7s" (OuterVolumeSpecName: "kube-api-access-qdx7s") pod "2d48f238-7fdf-4593-92e3-352f07e26062" (UID: "2d48f238-7fdf-4593-92e3-352f07e26062"). InnerVolumeSpecName "kube-api-access-qdx7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.547511 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d48f238-7fdf-4593-92e3-352f07e26062-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.547547 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdx7s\" (UniqueName: \"kubernetes.io/projected/2d48f238-7fdf-4593-92e3-352f07e26062-kube-api-access-qdx7s\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.791381 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jmvvv" Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.852510 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g92w\" (UniqueName: \"kubernetes.io/projected/5de674b4-317a-4d40-959c-6e0b050a18de-kube-api-access-6g92w\") pod \"5de674b4-317a-4d40-959c-6e0b050a18de\" (UID: \"5de674b4-317a-4d40-959c-6e0b050a18de\") " Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.852659 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5de674b4-317a-4d40-959c-6e0b050a18de-operator-scripts\") pod \"5de674b4-317a-4d40-959c-6e0b050a18de\" (UID: \"5de674b4-317a-4d40-959c-6e0b050a18de\") " Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.853746 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5de674b4-317a-4d40-959c-6e0b050a18de-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5de674b4-317a-4d40-959c-6e0b050a18de" (UID: "5de674b4-317a-4d40-959c-6e0b050a18de"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.856755 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de674b4-317a-4d40-959c-6e0b050a18de-kube-api-access-6g92w" (OuterVolumeSpecName: "kube-api-access-6g92w") pod "5de674b4-317a-4d40-959c-6e0b050a18de" (UID: "5de674b4-317a-4d40-959c-6e0b050a18de"). InnerVolumeSpecName "kube-api-access-6g92w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.955543 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g92w\" (UniqueName: \"kubernetes.io/projected/5de674b4-317a-4d40-959c-6e0b050a18de-kube-api-access-6g92w\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:02 crc kubenswrapper[4815]: I1205 09:30:02.955578 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5de674b4-317a-4d40-959c-6e0b050a18de-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.088235 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e03913c-e3a4-42e1-b133-5ef5f95d7b10","Type":"ContainerStarted","Data":"297c42c4df3d648eb7badc497ed93e52a142d1f33b5cae21b73a1cd8ed3de003"} Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.132063 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-41a2-account-create-update-rxfbj" event={"ID":"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c","Type":"ContainerDied","Data":"d99327d129745ab5888a16c30546eef5e75079e33f31f724ca719314b8c03678"} Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.132115 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d99327d129745ab5888a16c30546eef5e75079e33f31f724ca719314b8c03678" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.141154 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-41a2-account-create-update-rxfbj" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.151208 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" event={"ID":"2d48f238-7fdf-4593-92e3-352f07e26062","Type":"ContainerDied","Data":"70efe71fc22921e9ee5167ae6b525f493c32601e76a9936b1d23b7f03f3a815b"} Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.151245 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70efe71fc22921e9ee5167ae6b525f493c32601e76a9936b1d23b7f03f3a815b" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.151339 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.151441 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2bb0-account-create-update-f4zbp" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.157617 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qgqvs" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.157661 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-jmvvv" event={"ID":"5de674b4-317a-4d40-959c-6e0b050a18de","Type":"ContainerDied","Data":"41d3a8ebd821df59eb084c4bbb407faea8cfd222cc9aaa1e6768cc5beb59523a"} Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.157680 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-jmvvv" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.157692 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41d3a8ebd821df59eb084c4bbb407faea8cfd222cc9aaa1e6768cc5beb59523a" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.173898 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-88f59" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.263271 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0186806-095c-4433-bb68-55732c995bed-operator-scripts\") pod \"e0186806-095c-4433-bb68-55732c995bed\" (UID: \"e0186806-095c-4433-bb68-55732c995bed\") " Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.263377 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph4k4\" (UniqueName: \"kubernetes.io/projected/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-kube-api-access-ph4k4\") pod \"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c\" (UID: \"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c\") " Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.263451 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8rtr\" (UniqueName: \"kubernetes.io/projected/e0186806-095c-4433-bb68-55732c995bed-kube-api-access-l8rtr\") pod \"e0186806-095c-4433-bb68-55732c995bed\" (UID: \"e0186806-095c-4433-bb68-55732c995bed\") " Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.263469 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-operator-scripts\") pod \"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c\" (UID: \"b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c\") " Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.263514 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be43b898-0145-4ee3-aeb0-0e35acaa094d-operator-scripts\") pod \"be43b898-0145-4ee3-aeb0-0e35acaa094d\" (UID: \"be43b898-0145-4ee3-aeb0-0e35acaa094d\") " Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.263547 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6b4wq\" (UniqueName: \"kubernetes.io/projected/be43b898-0145-4ee3-aeb0-0e35acaa094d-kube-api-access-6b4wq\") pod \"be43b898-0145-4ee3-aeb0-0e35acaa094d\" (UID: \"be43b898-0145-4ee3-aeb0-0e35acaa094d\") " Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.263576 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnxvh\" (UniqueName: \"kubernetes.io/projected/15e7e2fa-915a-407e-a8cd-1a0225692722-kube-api-access-bnxvh\") pod \"15e7e2fa-915a-407e-a8cd-1a0225692722\" (UID: \"15e7e2fa-915a-407e-a8cd-1a0225692722\") " Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.263613 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15e7e2fa-915a-407e-a8cd-1a0225692722-operator-scripts\") pod \"15e7e2fa-915a-407e-a8cd-1a0225692722\" (UID: \"15e7e2fa-915a-407e-a8cd-1a0225692722\") " Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.264809 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15e7e2fa-915a-407e-a8cd-1a0225692722-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "15e7e2fa-915a-407e-a8cd-1a0225692722" (UID: "15e7e2fa-915a-407e-a8cd-1a0225692722"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.264970 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c" (UID: "b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.265294 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be43b898-0145-4ee3-aeb0-0e35acaa094d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "be43b898-0145-4ee3-aeb0-0e35acaa094d" (UID: "be43b898-0145-4ee3-aeb0-0e35acaa094d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.265328 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0186806-095c-4433-bb68-55732c995bed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e0186806-095c-4433-bb68-55732c995bed" (UID: "e0186806-095c-4433-bb68-55732c995bed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.270327 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-kube-api-access-ph4k4" (OuterVolumeSpecName: "kube-api-access-ph4k4") pod "b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c" (UID: "b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c"). InnerVolumeSpecName "kube-api-access-ph4k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.273686 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15e7e2fa-915a-407e-a8cd-1a0225692722-kube-api-access-bnxvh" (OuterVolumeSpecName: "kube-api-access-bnxvh") pod "15e7e2fa-915a-407e-a8cd-1a0225692722" (UID: "15e7e2fa-915a-407e-a8cd-1a0225692722"). InnerVolumeSpecName "kube-api-access-bnxvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.273781 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be43b898-0145-4ee3-aeb0-0e35acaa094d-kube-api-access-6b4wq" (OuterVolumeSpecName: "kube-api-access-6b4wq") pod "be43b898-0145-4ee3-aeb0-0e35acaa094d" (UID: "be43b898-0145-4ee3-aeb0-0e35acaa094d"). InnerVolumeSpecName "kube-api-access-6b4wq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.279560 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0186806-095c-4433-bb68-55732c995bed-kube-api-access-l8rtr" (OuterVolumeSpecName: "kube-api-access-l8rtr") pod "e0186806-095c-4433-bb68-55732c995bed" (UID: "e0186806-095c-4433-bb68-55732c995bed"). InnerVolumeSpecName "kube-api-access-l8rtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.366025 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnxvh\" (UniqueName: \"kubernetes.io/projected/15e7e2fa-915a-407e-a8cd-1a0225692722-kube-api-access-bnxvh\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.366064 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/15e7e2fa-915a-407e-a8cd-1a0225692722-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.366075 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e0186806-095c-4433-bb68-55732c995bed-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.366083 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph4k4\" (UniqueName: \"kubernetes.io/projected/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-kube-api-access-ph4k4\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.366092 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8rtr\" (UniqueName: \"kubernetes.io/projected/e0186806-095c-4433-bb68-55732c995bed-kube-api-access-l8rtr\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.366100 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.366112 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be43b898-0145-4ee3-aeb0-0e35acaa094d-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.366119 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6b4wq\" (UniqueName: \"kubernetes.io/projected/be43b898-0145-4ee3-aeb0-0e35acaa094d-kube-api-access-6b4wq\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.732344 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.791142 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwbhs\" (UniqueName: \"kubernetes.io/projected/0b2189a3-9b89-4167-91ce-b9e92a360d11-kube-api-access-kwbhs\") pod \"0b2189a3-9b89-4167-91ce-b9e92a360d11\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.791248 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b2189a3-9b89-4167-91ce-b9e92a360d11-secret-volume\") pod \"0b2189a3-9b89-4167-91ce-b9e92a360d11\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.791328 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b2189a3-9b89-4167-91ce-b9e92a360d11-config-volume\") pod \"0b2189a3-9b89-4167-91ce-b9e92a360d11\" (UID: \"0b2189a3-9b89-4167-91ce-b9e92a360d11\") " Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.798784 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b2189a3-9b89-4167-91ce-b9e92a360d11-kube-api-access-kwbhs" (OuterVolumeSpecName: "kube-api-access-kwbhs") pod "0b2189a3-9b89-4167-91ce-b9e92a360d11" (UID: "0b2189a3-9b89-4167-91ce-b9e92a360d11"). InnerVolumeSpecName "kube-api-access-kwbhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.801906 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b2189a3-9b89-4167-91ce-b9e92a360d11-config-volume" (OuterVolumeSpecName: "config-volume") pod "0b2189a3-9b89-4167-91ce-b9e92a360d11" (UID: "0b2189a3-9b89-4167-91ce-b9e92a360d11"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.807793 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwbhs\" (UniqueName: \"kubernetes.io/projected/0b2189a3-9b89-4167-91ce-b9e92a360d11-kube-api-access-kwbhs\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.807843 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/0b2189a3-9b89-4167-91ce-b9e92a360d11-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.818606 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b2189a3-9b89-4167-91ce-b9e92a360d11-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "0b2189a3-9b89-4167-91ce-b9e92a360d11" (UID: "0b2189a3-9b89-4167-91ce-b9e92a360d11"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:03 crc kubenswrapper[4815]: I1205 09:30:03.909186 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/0b2189a3-9b89-4167-91ce-b9e92a360d11-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.166941 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-qgqvs" event={"ID":"be43b898-0145-4ee3-aeb0-0e35acaa094d","Type":"ContainerDied","Data":"86322404e62f0b6ad6046cd3bfe83b1b961519b8bd0c9f90fce852b9a46ffbd4"} Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.167225 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="86322404e62f0b6ad6046cd3bfe83b1b961519b8bd0c9f90fce852b9a46ffbd4" Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.167318 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-qgqvs" Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.169946 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-88f59" event={"ID":"e0186806-095c-4433-bb68-55732c995bed","Type":"ContainerDied","Data":"7191108af0eb96c1520a648d1824d816dc9fd40d1f3e779770a805164faa1060"} Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.169972 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7191108af0eb96c1520a648d1824d816dc9fd40d1f3e779770a805164faa1060" Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.170036 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-88f59" Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.172254 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" event={"ID":"15e7e2fa-915a-407e-a8cd-1a0225692722","Type":"ContainerDied","Data":"6c4d79fa7328539efe6954bcf59e9b5f6248c0007449f08a17fddd519d0c0169"} Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.172283 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c4d79fa7328539efe6954bcf59e9b5f6248c0007449f08a17fddd519d0c0169" Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.172337 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-d1d3-account-create-update-5gxdr" Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.174438 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-41a2-account-create-update-rxfbj" Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.174501 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" event={"ID":"0b2189a3-9b89-4167-91ce-b9e92a360d11","Type":"ContainerDied","Data":"57051021e529691c600359680419276c5b6b766567beef2898b57a9dd9745b24"} Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.174524 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57051021e529691c600359680419276c5b6b766567beef2898b57a9dd9745b24" Dec 05 09:30:04 crc kubenswrapper[4815]: I1205 09:30:04.174624 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8" Dec 05 09:30:05 crc kubenswrapper[4815]: I1205 09:30:05.186453 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e03913c-e3a4-42e1-b133-5ef5f95d7b10","Type":"ContainerStarted","Data":"f3388f69e8652b0f560c12981de679df22d6cf1dbce24ccbb34dc0314322a52b"} Dec 05 09:30:05 crc kubenswrapper[4815]: I1205 09:30:05.186527 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e03913c-e3a4-42e1-b133-5ef5f95d7b10","Type":"ContainerStarted","Data":"8011f7bff74e52e9fb6d190a39fa8c5709700fc614542e3dd4077d26ba85228c"} Dec 05 09:30:07 crc kubenswrapper[4815]: I1205 09:30:07.203943 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e03913c-e3a4-42e1-b133-5ef5f95d7b10","Type":"ContainerStarted","Data":"4c6fbd823baa707eca219357caadb96f1386f628ccfa02cd8cbe38810c2cc360"} Dec 05 09:30:07 crc kubenswrapper[4815]: I1205 09:30:07.204551 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:30:07 crc kubenswrapper[4815]: I1205 09:30:07.250988 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.582505191 podStartE2EDuration="7.250954002s" podCreationTimestamp="2025-12-05 09:30:00 +0000 UTC" firstStartedPulling="2025-12-05 09:30:01.642724794 +0000 UTC m=+1400.521331631" lastFinishedPulling="2025-12-05 09:30:06.311173605 +0000 UTC m=+1405.189780442" observedRunningTime="2025-12-05 09:30:07.249581054 +0000 UTC m=+1406.128187901" watchObservedRunningTime="2025-12-05 09:30:07.250954002 +0000 UTC m=+1406.129560829" Dec 05 09:30:07 crc kubenswrapper[4815]: I1205 09:30:07.956050 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.101178 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.184459 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-config\") pod \"7126486e-29e9-487c-b603-7bf23251591b\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.184648 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-ovndb-tls-certs\") pod \"7126486e-29e9-487c-b603-7bf23251591b\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.184780 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-combined-ca-bundle\") pod \"7126486e-29e9-487c-b603-7bf23251591b\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.184886 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8sx7l\" (UniqueName: \"kubernetes.io/projected/7126486e-29e9-487c-b603-7bf23251591b-kube-api-access-8sx7l\") pod \"7126486e-29e9-487c-b603-7bf23251591b\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.184943 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-httpd-config\") pod \"7126486e-29e9-487c-b603-7bf23251591b\" (UID: \"7126486e-29e9-487c-b603-7bf23251591b\") " Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.207855 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "7126486e-29e9-487c-b603-7bf23251591b" (UID: "7126486e-29e9-487c-b603-7bf23251591b"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.238707 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7126486e-29e9-487c-b603-7bf23251591b-kube-api-access-8sx7l" (OuterVolumeSpecName: "kube-api-access-8sx7l") pod "7126486e-29e9-487c-b603-7bf23251591b" (UID: "7126486e-29e9-487c-b603-7bf23251591b"). InnerVolumeSpecName "kube-api-access-8sx7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.270594 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s5qnf"] Dec 05 09:30:08 crc kubenswrapper[4815]: E1205 09:30:08.271033 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0186806-095c-4433-bb68-55732c995bed" containerName="mariadb-database-create" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271054 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0186806-095c-4433-bb68-55732c995bed" containerName="mariadb-database-create" Dec 05 09:30:08 crc kubenswrapper[4815]: E1205 09:30:08.271069 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be43b898-0145-4ee3-aeb0-0e35acaa094d" containerName="mariadb-database-create" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271075 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="be43b898-0145-4ee3-aeb0-0e35acaa094d" containerName="mariadb-database-create" Dec 05 09:30:08 crc kubenswrapper[4815]: E1205 09:30:08.271106 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2189a3-9b89-4167-91ce-b9e92a360d11" containerName="collect-profiles" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271113 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2189a3-9b89-4167-91ce-b9e92a360d11" containerName="collect-profiles" Dec 05 09:30:08 crc kubenswrapper[4815]: E1205 09:30:08.271128 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7126486e-29e9-487c-b603-7bf23251591b" containerName="neutron-httpd" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271134 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7126486e-29e9-487c-b603-7bf23251591b" containerName="neutron-httpd" Dec 05 09:30:08 crc kubenswrapper[4815]: E1205 09:30:08.271144 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c" containerName="mariadb-account-create-update" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271150 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c" containerName="mariadb-account-create-update" Dec 05 09:30:08 crc kubenswrapper[4815]: E1205 09:30:08.271160 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7126486e-29e9-487c-b603-7bf23251591b" containerName="neutron-api" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271166 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7126486e-29e9-487c-b603-7bf23251591b" containerName="neutron-api" Dec 05 09:30:08 crc kubenswrapper[4815]: E1205 09:30:08.271182 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15e7e2fa-915a-407e-a8cd-1a0225692722" containerName="mariadb-account-create-update" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271188 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="15e7e2fa-915a-407e-a8cd-1a0225692722" containerName="mariadb-account-create-update" Dec 05 09:30:08 crc kubenswrapper[4815]: E1205 09:30:08.271195 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d48f238-7fdf-4593-92e3-352f07e26062" containerName="mariadb-account-create-update" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271201 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d48f238-7fdf-4593-92e3-352f07e26062" containerName="mariadb-account-create-update" Dec 05 09:30:08 crc kubenswrapper[4815]: E1205 09:30:08.271209 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de674b4-317a-4d40-959c-6e0b050a18de" containerName="mariadb-database-create" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271214 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de674b4-317a-4d40-959c-6e0b050a18de" containerName="mariadb-database-create" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271518 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d48f238-7fdf-4593-92e3-352f07e26062" containerName="mariadb-account-create-update" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271540 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0186806-095c-4433-bb68-55732c995bed" containerName="mariadb-database-create" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271609 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7126486e-29e9-487c-b603-7bf23251591b" containerName="neutron-api" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271624 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de674b4-317a-4d40-959c-6e0b050a18de" containerName="mariadb-database-create" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271633 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="15e7e2fa-915a-407e-a8cd-1a0225692722" containerName="mariadb-account-create-update" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271640 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7126486e-29e9-487c-b603-7bf23251591b" containerName="neutron-httpd" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271703 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c" containerName="mariadb-account-create-update" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271712 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2189a3-9b89-4167-91ce-b9e92a360d11" containerName="collect-profiles" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.271742 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="be43b898-0145-4ee3-aeb0-0e35acaa094d" containerName="mariadb-database-create" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.272418 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.280157 4815 generic.go:334] "Generic (PLEG): container finished" podID="7126486e-29e9-487c-b603-7bf23251591b" containerID="697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab" exitCode=0 Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.286168 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6bdb56b95d-k5pvr" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.286365 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdb56b95d-k5pvr" event={"ID":"7126486e-29e9-487c-b603-7bf23251591b","Type":"ContainerDied","Data":"697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab"} Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.286407 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6bdb56b95d-k5pvr" event={"ID":"7126486e-29e9-487c-b603-7bf23251591b","Type":"ContainerDied","Data":"1bf4df7c20ea4af99bcb76413a080170c1d6096dca3b00ab23731d5711af0964"} Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.286456 4815 scope.go:117] "RemoveContainer" containerID="9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.286936 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8sx7l\" (UniqueName: \"kubernetes.io/projected/7126486e-29e9-487c-b603-7bf23251591b-kube-api-access-8sx7l\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.286966 4815 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.287257 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.287433 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.300004 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-qklpl" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.326880 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s5qnf"] Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.370097 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7126486e-29e9-487c-b603-7bf23251591b" (UID: "7126486e-29e9-487c-b603-7bf23251591b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.372710 4815 scope.go:117] "RemoveContainer" containerID="697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.378218 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-config" (OuterVolumeSpecName: "config") pod "7126486e-29e9-487c-b603-7bf23251591b" (UID: "7126486e-29e9-487c-b603-7bf23251591b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.398947 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dfgr\" (UniqueName: \"kubernetes.io/projected/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-kube-api-access-2dfgr\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.399014 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.399040 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-config-data\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.399086 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-scripts\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.399252 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.399265 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.404953 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "7126486e-29e9-487c-b603-7bf23251591b" (UID: "7126486e-29e9-487c-b603-7bf23251591b"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.418437 4815 scope.go:117] "RemoveContainer" containerID="9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35" Dec 05 09:30:08 crc kubenswrapper[4815]: E1205 09:30:08.419059 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35\": container with ID starting with 9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35 not found: ID does not exist" containerID="9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.419108 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35"} err="failed to get container status \"9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35\": rpc error: code = NotFound desc = could not find container \"9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35\": container with ID starting with 9ae14dae3c27a52c048be8bc584602c22477c4d8d9c3a21ed772b7d13bcecf35 not found: ID does not exist" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.419159 4815 scope.go:117] "RemoveContainer" containerID="697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab" Dec 05 09:30:08 crc kubenswrapper[4815]: E1205 09:30:08.420255 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab\": container with ID starting with 697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab not found: ID does not exist" containerID="697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.420286 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab"} err="failed to get container status \"697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab\": rpc error: code = NotFound desc = could not find container \"697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab\": container with ID starting with 697562f3a5696d617fe87a6553ffcd0fe52487d759ed895548d5778616356eab not found: ID does not exist" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.500593 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-scripts\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.500733 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dfgr\" (UniqueName: \"kubernetes.io/projected/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-kube-api-access-2dfgr\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.500790 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.500837 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-config-data\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.501469 4815 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7126486e-29e9-487c-b603-7bf23251591b-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.504230 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-scripts\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.504474 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.507694 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-config-data\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.519990 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dfgr\" (UniqueName: \"kubernetes.io/projected/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-kube-api-access-2dfgr\") pod \"nova-cell0-conductor-db-sync-s5qnf\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.621725 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6bdb56b95d-k5pvr"] Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.629016 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6bdb56b95d-k5pvr"] Dec 05 09:30:08 crc kubenswrapper[4815]: I1205 09:30:08.675305 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:09 crc kubenswrapper[4815]: I1205 09:30:09.217919 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s5qnf"] Dec 05 09:30:09 crc kubenswrapper[4815]: I1205 09:30:09.294146 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="ceilometer-central-agent" containerID="cri-o://297c42c4df3d648eb7badc497ed93e52a142d1f33b5cae21b73a1cd8ed3de003" gracePeriod=30 Dec 05 09:30:09 crc kubenswrapper[4815]: I1205 09:30:09.294250 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s5qnf" event={"ID":"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4","Type":"ContainerStarted","Data":"ce3f55feafa9f31a392e757f5541eaf6b8f04e9ff15deee6788de3c8c97b0b7d"} Dec 05 09:30:09 crc kubenswrapper[4815]: I1205 09:30:09.294687 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="proxy-httpd" containerID="cri-o://4c6fbd823baa707eca219357caadb96f1386f628ccfa02cd8cbe38810c2cc360" gracePeriod=30 Dec 05 09:30:09 crc kubenswrapper[4815]: I1205 09:30:09.294758 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="sg-core" containerID="cri-o://f3388f69e8652b0f560c12981de679df22d6cf1dbce24ccbb34dc0314322a52b" gracePeriod=30 Dec 05 09:30:09 crc kubenswrapper[4815]: I1205 09:30:09.294807 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="ceilometer-notification-agent" containerID="cri-o://8011f7bff74e52e9fb6d190a39fa8c5709700fc614542e3dd4077d26ba85228c" gracePeriod=30 Dec 05 09:30:09 crc kubenswrapper[4815]: I1205 09:30:09.434556 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7126486e-29e9-487c-b603-7bf23251591b" path="/var/lib/kubelet/pods/7126486e-29e9-487c-b603-7bf23251591b/volumes" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.305698 4815 generic.go:334] "Generic (PLEG): container finished" podID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerID="4c6fbd823baa707eca219357caadb96f1386f628ccfa02cd8cbe38810c2cc360" exitCode=0 Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.305972 4815 generic.go:334] "Generic (PLEG): container finished" podID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerID="f3388f69e8652b0f560c12981de679df22d6cf1dbce24ccbb34dc0314322a52b" exitCode=2 Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.305982 4815 generic.go:334] "Generic (PLEG): container finished" podID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerID="8011f7bff74e52e9fb6d190a39fa8c5709700fc614542e3dd4077d26ba85228c" exitCode=0 Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.305990 4815 generic.go:334] "Generic (PLEG): container finished" podID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerID="297c42c4df3d648eb7badc497ed93e52a142d1f33b5cae21b73a1cd8ed3de003" exitCode=0 Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.305804 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e03913c-e3a4-42e1-b133-5ef5f95d7b10","Type":"ContainerDied","Data":"4c6fbd823baa707eca219357caadb96f1386f628ccfa02cd8cbe38810c2cc360"} Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.306025 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e03913c-e3a4-42e1-b133-5ef5f95d7b10","Type":"ContainerDied","Data":"f3388f69e8652b0f560c12981de679df22d6cf1dbce24ccbb34dc0314322a52b"} Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.306038 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e03913c-e3a4-42e1-b133-5ef5f95d7b10","Type":"ContainerDied","Data":"8011f7bff74e52e9fb6d190a39fa8c5709700fc614542e3dd4077d26ba85228c"} Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.306048 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e03913c-e3a4-42e1-b133-5ef5f95d7b10","Type":"ContainerDied","Data":"297c42c4df3d648eb7badc497ed93e52a142d1f33b5cae21b73a1cd8ed3de003"} Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.306056 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7e03913c-e3a4-42e1-b133-5ef5f95d7b10","Type":"ContainerDied","Data":"0d93a1fc85ac4130a6fa340c246b3caf51a93e2f746cb15228dc90c7f94e2238"} Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.306066 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d93a1fc85ac4130a6fa340c246b3caf51a93e2f746cb15228dc90c7f94e2238" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.322692 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.451404 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6p5z\" (UniqueName: \"kubernetes.io/projected/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-kube-api-access-p6p5z\") pod \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.451466 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-log-httpd\") pod \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.451509 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-combined-ca-bundle\") pod \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.451552 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-scripts\") pod \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.451682 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-config-data\") pod \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.451717 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-sg-core-conf-yaml\") pod \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.452461 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-run-httpd\") pod \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\" (UID: \"7e03913c-e3a4-42e1-b133-5ef5f95d7b10\") " Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.452671 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7e03913c-e3a4-42e1-b133-5ef5f95d7b10" (UID: "7e03913c-e3a4-42e1-b133-5ef5f95d7b10"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.452715 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7e03913c-e3a4-42e1-b133-5ef5f95d7b10" (UID: "7e03913c-e3a4-42e1-b133-5ef5f95d7b10"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.452967 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.452982 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.458242 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-scripts" (OuterVolumeSpecName: "scripts") pod "7e03913c-e3a4-42e1-b133-5ef5f95d7b10" (UID: "7e03913c-e3a4-42e1-b133-5ef5f95d7b10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.465599 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-kube-api-access-p6p5z" (OuterVolumeSpecName: "kube-api-access-p6p5z") pod "7e03913c-e3a4-42e1-b133-5ef5f95d7b10" (UID: "7e03913c-e3a4-42e1-b133-5ef5f95d7b10"). InnerVolumeSpecName "kube-api-access-p6p5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.490242 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7e03913c-e3a4-42e1-b133-5ef5f95d7b10" (UID: "7e03913c-e3a4-42e1-b133-5ef5f95d7b10"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.548901 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7e03913c-e3a4-42e1-b133-5ef5f95d7b10" (UID: "7e03913c-e3a4-42e1-b133-5ef5f95d7b10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.554590 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.554612 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6p5z\" (UniqueName: \"kubernetes.io/projected/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-kube-api-access-p6p5z\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.554621 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.554629 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.605571 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-config-data" (OuterVolumeSpecName: "config-data") pod "7e03913c-e3a4-42e1-b133-5ef5f95d7b10" (UID: "7e03913c-e3a4-42e1-b133-5ef5f95d7b10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:10 crc kubenswrapper[4815]: I1205 09:30:10.656720 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7e03913c-e3a4-42e1-b133-5ef5f95d7b10-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.327035 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.385786 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.405335 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.421672 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:11 crc kubenswrapper[4815]: E1205 09:30:11.422010 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="sg-core" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.422023 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="sg-core" Dec 05 09:30:11 crc kubenswrapper[4815]: E1205 09:30:11.422043 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="ceilometer-central-agent" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.422049 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="ceilometer-central-agent" Dec 05 09:30:11 crc kubenswrapper[4815]: E1205 09:30:11.422058 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="ceilometer-notification-agent" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.422064 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="ceilometer-notification-agent" Dec 05 09:30:11 crc kubenswrapper[4815]: E1205 09:30:11.422074 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="proxy-httpd" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.422079 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="proxy-httpd" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.422233 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="proxy-httpd" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.422245 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="sg-core" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.422258 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="ceilometer-central-agent" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.422270 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" containerName="ceilometer-notification-agent" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.431135 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.441286 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.441442 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.441635 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e03913c-e3a4-42e1-b133-5ef5f95d7b10" path="/var/lib/kubelet/pods/7e03913c-e3a4-42e1-b133-5ef5f95d7b10/volumes" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.464056 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.574400 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw7bm\" (UniqueName: \"kubernetes.io/projected/1356a8ab-0838-4bfa-9e4b-f90ad6714539-kube-api-access-sw7bm\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.574462 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.574504 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-run-httpd\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.574529 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-config-data\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.574594 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-log-httpd\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.574632 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-scripts\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.574672 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.676330 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.676443 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw7bm\" (UniqueName: \"kubernetes.io/projected/1356a8ab-0838-4bfa-9e4b-f90ad6714539-kube-api-access-sw7bm\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.676471 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.676513 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-run-httpd\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.676535 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-config-data\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.676579 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-log-httpd\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.676605 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-scripts\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.678256 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-run-httpd\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.678523 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-log-httpd\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.681858 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.693391 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-config-data\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.693815 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-scripts\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.703841 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.704082 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw7bm\" (UniqueName: \"kubernetes.io/projected/1356a8ab-0838-4bfa-9e4b-f90ad6714539-kube-api-access-sw7bm\") pod \"ceilometer-0\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " pod="openstack/ceilometer-0" Dec 05 09:30:11 crc kubenswrapper[4815]: I1205 09:30:11.773533 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:12 crc kubenswrapper[4815]: I1205 09:30:12.078875 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:12 crc kubenswrapper[4815]: W1205 09:30:12.087833 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1356a8ab_0838_4bfa_9e4b_f90ad6714539.slice/crio-128a5e85123762126383cad639c7be351146be55686e45210bfd13eeb81878cc WatchSource:0}: Error finding container 128a5e85123762126383cad639c7be351146be55686e45210bfd13eeb81878cc: Status 404 returned error can't find the container with id 128a5e85123762126383cad639c7be351146be55686e45210bfd13eeb81878cc Dec 05 09:30:12 crc kubenswrapper[4815]: I1205 09:30:12.355409 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1356a8ab-0838-4bfa-9e4b-f90ad6714539","Type":"ContainerStarted","Data":"128a5e85123762126383cad639c7be351146be55686e45210bfd13eeb81878cc"} Dec 05 09:30:12 crc kubenswrapper[4815]: I1205 09:30:12.505818 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:13 crc kubenswrapper[4815]: I1205 09:30:13.369190 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1356a8ab-0838-4bfa-9e4b-f90ad6714539","Type":"ContainerStarted","Data":"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e"} Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.192116 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.192476 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.192547 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.193310 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ca66886398a56fe8786c896f7db80c714344ffc63db0667dbec913c482105f33"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.193655 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://ca66886398a56fe8786c896f7db80c714344ffc63db0667dbec913c482105f33" gracePeriod=600 Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.479640 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s5qnf" event={"ID":"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4","Type":"ContainerStarted","Data":"ffe9faeaa0bf8cd09661f677d909031160c5256658cc3eb14e3506e73ac4bd46"} Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.486280 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="ca66886398a56fe8786c896f7db80c714344ffc63db0667dbec913c482105f33" exitCode=0 Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.486356 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"ca66886398a56fe8786c896f7db80c714344ffc63db0667dbec913c482105f33"} Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.486401 4815 scope.go:117] "RemoveContainer" containerID="e422e582a96b96c6d83e3c0fc7b930a3d237267c29a704a2f4827be174a6d09c" Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.489548 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1356a8ab-0838-4bfa-9e4b-f90ad6714539","Type":"ContainerStarted","Data":"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566"} Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.489587 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1356a8ab-0838-4bfa-9e4b-f90ad6714539","Type":"ContainerStarted","Data":"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34"} Dec 05 09:30:20 crc kubenswrapper[4815]: I1205 09:30:20.526305 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-s5qnf" podStartSLOduration=2.283665916 podStartE2EDuration="12.526287222s" podCreationTimestamp="2025-12-05 09:30:08 +0000 UTC" firstStartedPulling="2025-12-05 09:30:09.233980694 +0000 UTC m=+1408.112587531" lastFinishedPulling="2025-12-05 09:30:19.476602 +0000 UTC m=+1418.355208837" observedRunningTime="2025-12-05 09:30:20.516030901 +0000 UTC m=+1419.394637758" watchObservedRunningTime="2025-12-05 09:30:20.526287222 +0000 UTC m=+1419.404894059" Dec 05 09:30:21 crc kubenswrapper[4815]: I1205 09:30:21.500425 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa"} Dec 05 09:30:23 crc kubenswrapper[4815]: I1205 09:30:23.525181 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1356a8ab-0838-4bfa-9e4b-f90ad6714539","Type":"ContainerStarted","Data":"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785"} Dec 05 09:30:23 crc kubenswrapper[4815]: I1205 09:30:23.525841 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="ceilometer-central-agent" containerID="cri-o://cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e" gracePeriod=30 Dec 05 09:30:23 crc kubenswrapper[4815]: I1205 09:30:23.526151 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:30:23 crc kubenswrapper[4815]: I1205 09:30:23.526195 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="ceilometer-notification-agent" containerID="cri-o://e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34" gracePeriod=30 Dec 05 09:30:23 crc kubenswrapper[4815]: I1205 09:30:23.526197 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="sg-core" containerID="cri-o://3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566" gracePeriod=30 Dec 05 09:30:23 crc kubenswrapper[4815]: I1205 09:30:23.526268 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="proxy-httpd" containerID="cri-o://addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785" gracePeriod=30 Dec 05 09:30:23 crc kubenswrapper[4815]: I1205 09:30:23.576252 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.656171707 podStartE2EDuration="12.576225618s" podCreationTimestamp="2025-12-05 09:30:11 +0000 UTC" firstStartedPulling="2025-12-05 09:30:12.0933624 +0000 UTC m=+1410.971969237" lastFinishedPulling="2025-12-05 09:30:23.013416311 +0000 UTC m=+1421.892023148" observedRunningTime="2025-12-05 09:30:23.564208559 +0000 UTC m=+1422.442815396" watchObservedRunningTime="2025-12-05 09:30:23.576225618 +0000 UTC m=+1422.454832445" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.388554 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.440555 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-sg-core-conf-yaml\") pod \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.440626 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-run-httpd\") pod \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.440737 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw7bm\" (UniqueName: \"kubernetes.io/projected/1356a8ab-0838-4bfa-9e4b-f90ad6714539-kube-api-access-sw7bm\") pod \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.440784 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-combined-ca-bundle\") pod \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.440900 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-config-data\") pod \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.440921 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-scripts\") pod \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.440986 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-log-httpd\") pod \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\" (UID: \"1356a8ab-0838-4bfa-9e4b-f90ad6714539\") " Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.441700 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1356a8ab-0838-4bfa-9e4b-f90ad6714539" (UID: "1356a8ab-0838-4bfa-9e4b-f90ad6714539"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.441871 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1356a8ab-0838-4bfa-9e4b-f90ad6714539" (UID: "1356a8ab-0838-4bfa-9e4b-f90ad6714539"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.448699 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1356a8ab-0838-4bfa-9e4b-f90ad6714539-kube-api-access-sw7bm" (OuterVolumeSpecName: "kube-api-access-sw7bm") pod "1356a8ab-0838-4bfa-9e4b-f90ad6714539" (UID: "1356a8ab-0838-4bfa-9e4b-f90ad6714539"). InnerVolumeSpecName "kube-api-access-sw7bm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.451697 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-scripts" (OuterVolumeSpecName: "scripts") pod "1356a8ab-0838-4bfa-9e4b-f90ad6714539" (UID: "1356a8ab-0838-4bfa-9e4b-f90ad6714539"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.470993 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1356a8ab-0838-4bfa-9e4b-f90ad6714539" (UID: "1356a8ab-0838-4bfa-9e4b-f90ad6714539"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.539558 4815 generic.go:334] "Generic (PLEG): container finished" podID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerID="addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785" exitCode=0 Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.539614 4815 generic.go:334] "Generic (PLEG): container finished" podID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerID="3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566" exitCode=2 Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.539623 4815 generic.go:334] "Generic (PLEG): container finished" podID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerID="e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34" exitCode=0 Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.539630 4815 generic.go:334] "Generic (PLEG): container finished" podID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerID="cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e" exitCode=0 Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.539654 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1356a8ab-0838-4bfa-9e4b-f90ad6714539","Type":"ContainerDied","Data":"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785"} Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.539685 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1356a8ab-0838-4bfa-9e4b-f90ad6714539","Type":"ContainerDied","Data":"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566"} Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.539696 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1356a8ab-0838-4bfa-9e4b-f90ad6714539","Type":"ContainerDied","Data":"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34"} Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.539709 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1356a8ab-0838-4bfa-9e4b-f90ad6714539","Type":"ContainerDied","Data":"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e"} Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.539756 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1356a8ab-0838-4bfa-9e4b-f90ad6714539","Type":"ContainerDied","Data":"128a5e85123762126383cad639c7be351146be55686e45210bfd13eeb81878cc"} Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.539774 4815 scope.go:117] "RemoveContainer" containerID="addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.539985 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.543081 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.543550 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.543655 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.543751 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1356a8ab-0838-4bfa-9e4b-f90ad6714539-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.543859 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw7bm\" (UniqueName: \"kubernetes.io/projected/1356a8ab-0838-4bfa-9e4b-f90ad6714539-kube-api-access-sw7bm\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.555163 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1356a8ab-0838-4bfa-9e4b-f90ad6714539" (UID: "1356a8ab-0838-4bfa-9e4b-f90ad6714539"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.563263 4815 scope.go:117] "RemoveContainer" containerID="3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.594147 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-config-data" (OuterVolumeSpecName: "config-data") pod "1356a8ab-0838-4bfa-9e4b-f90ad6714539" (UID: "1356a8ab-0838-4bfa-9e4b-f90ad6714539"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.601581 4815 scope.go:117] "RemoveContainer" containerID="e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.629599 4815 scope.go:117] "RemoveContainer" containerID="cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.645232 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.645268 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1356a8ab-0838-4bfa-9e4b-f90ad6714539-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.649404 4815 scope.go:117] "RemoveContainer" containerID="addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785" Dec 05 09:30:24 crc kubenswrapper[4815]: E1205 09:30:24.650416 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785\": container with ID starting with addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785 not found: ID does not exist" containerID="addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.650450 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785"} err="failed to get container status \"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785\": rpc error: code = NotFound desc = could not find container \"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785\": container with ID starting with addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.650472 4815 scope.go:117] "RemoveContainer" containerID="3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566" Dec 05 09:30:24 crc kubenswrapper[4815]: E1205 09:30:24.650942 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566\": container with ID starting with 3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566 not found: ID does not exist" containerID="3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.650961 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566"} err="failed to get container status \"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566\": rpc error: code = NotFound desc = could not find container \"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566\": container with ID starting with 3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.650973 4815 scope.go:117] "RemoveContainer" containerID="e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34" Dec 05 09:30:24 crc kubenswrapper[4815]: E1205 09:30:24.651699 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34\": container with ID starting with e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34 not found: ID does not exist" containerID="e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.651722 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34"} err="failed to get container status \"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34\": rpc error: code = NotFound desc = could not find container \"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34\": container with ID starting with e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.651735 4815 scope.go:117] "RemoveContainer" containerID="cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e" Dec 05 09:30:24 crc kubenswrapper[4815]: E1205 09:30:24.651993 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e\": container with ID starting with cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e not found: ID does not exist" containerID="cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.652011 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e"} err="failed to get container status \"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e\": rpc error: code = NotFound desc = could not find container \"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e\": container with ID starting with cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.652024 4815 scope.go:117] "RemoveContainer" containerID="addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.652257 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785"} err="failed to get container status \"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785\": rpc error: code = NotFound desc = could not find container \"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785\": container with ID starting with addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.652282 4815 scope.go:117] "RemoveContainer" containerID="3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.652585 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566"} err="failed to get container status \"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566\": rpc error: code = NotFound desc = could not find container \"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566\": container with ID starting with 3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.652607 4815 scope.go:117] "RemoveContainer" containerID="e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.652877 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34"} err="failed to get container status \"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34\": rpc error: code = NotFound desc = could not find container \"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34\": container with ID starting with e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.652896 4815 scope.go:117] "RemoveContainer" containerID="cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.653142 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e"} err="failed to get container status \"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e\": rpc error: code = NotFound desc = could not find container \"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e\": container with ID starting with cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.653159 4815 scope.go:117] "RemoveContainer" containerID="addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.653576 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785"} err="failed to get container status \"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785\": rpc error: code = NotFound desc = could not find container \"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785\": container with ID starting with addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.653597 4815 scope.go:117] "RemoveContainer" containerID="3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.653834 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566"} err="failed to get container status \"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566\": rpc error: code = NotFound desc = could not find container \"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566\": container with ID starting with 3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.653864 4815 scope.go:117] "RemoveContainer" containerID="e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.654165 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34"} err="failed to get container status \"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34\": rpc error: code = NotFound desc = could not find container \"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34\": container with ID starting with e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.654183 4815 scope.go:117] "RemoveContainer" containerID="cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.654562 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e"} err="failed to get container status \"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e\": rpc error: code = NotFound desc = could not find container \"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e\": container with ID starting with cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.654581 4815 scope.go:117] "RemoveContainer" containerID="addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.654816 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785"} err="failed to get container status \"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785\": rpc error: code = NotFound desc = could not find container \"addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785\": container with ID starting with addec2e79816091a3186002af9b3e50d54266c10e23ae06ec47661f5e80b8785 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.654834 4815 scope.go:117] "RemoveContainer" containerID="3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.655073 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566"} err="failed to get container status \"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566\": rpc error: code = NotFound desc = could not find container \"3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566\": container with ID starting with 3f1008ac94420edee0a5bfd92d1b4f8c3c14b92c1ad6ca039909648bfd8c9566 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.655091 4815 scope.go:117] "RemoveContainer" containerID="e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.655695 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34"} err="failed to get container status \"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34\": rpc error: code = NotFound desc = could not find container \"e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34\": container with ID starting with e62325a602fa6b287490cb4103547c6d5641c9ec2314f6a193092577516fca34 not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.655713 4815 scope.go:117] "RemoveContainer" containerID="cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.656275 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e"} err="failed to get container status \"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e\": rpc error: code = NotFound desc = could not find container \"cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e\": container with ID starting with cd7243145952d4736053020b28abfa95571c7fb5ba86db69cf1d38580d946e2e not found: ID does not exist" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.873051 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.882819 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.899349 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:24 crc kubenswrapper[4815]: E1205 09:30:24.899770 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="proxy-httpd" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.899791 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="proxy-httpd" Dec 05 09:30:24 crc kubenswrapper[4815]: E1205 09:30:24.899805 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="ceilometer-notification-agent" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.899811 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="ceilometer-notification-agent" Dec 05 09:30:24 crc kubenswrapper[4815]: E1205 09:30:24.899824 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="ceilometer-central-agent" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.899831 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="ceilometer-central-agent" Dec 05 09:30:24 crc kubenswrapper[4815]: E1205 09:30:24.899854 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="sg-core" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.899860 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="sg-core" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.900017 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="ceilometer-notification-agent" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.900027 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="ceilometer-central-agent" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.900039 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="proxy-httpd" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.900049 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" containerName="sg-core" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.901531 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.903208 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.905908 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.921655 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.949437 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-config-data\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.949536 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-scripts\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.949566 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mkm5\" (UniqueName: \"kubernetes.io/projected/fcf9574d-dedc-460f-9799-f556d817fb52-kube-api-access-6mkm5\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.949615 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-log-httpd\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.949678 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.949730 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:24 crc kubenswrapper[4815]: I1205 09:30:24.949755 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-run-httpd\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.050996 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-config-data\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.051237 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-scripts\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.051324 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mkm5\" (UniqueName: \"kubernetes.io/projected/fcf9574d-dedc-460f-9799-f556d817fb52-kube-api-access-6mkm5\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.051460 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-log-httpd\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.051587 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.051699 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.051804 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-run-httpd\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.052156 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-log-httpd\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.052360 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-run-httpd\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.055733 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-config-data\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.058106 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-scripts\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.059150 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.068878 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mkm5\" (UniqueName: \"kubernetes.io/projected/fcf9574d-dedc-460f-9799-f556d817fb52-kube-api-access-6mkm5\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.069337 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.230920 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.465759 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1356a8ab-0838-4bfa-9e4b-f90ad6714539" path="/var/lib/kubelet/pods/1356a8ab-0838-4bfa-9e4b-f90ad6714539/volumes" Dec 05 09:30:25 crc kubenswrapper[4815]: W1205 09:30:25.725449 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfcf9574d_dedc_460f_9799_f556d817fb52.slice/crio-6616ac294244eaa9db5655bc3c3908deb54ac21d4d848caa064ae2774800d360 WatchSource:0}: Error finding container 6616ac294244eaa9db5655bc3c3908deb54ac21d4d848caa064ae2774800d360: Status 404 returned error can't find the container with id 6616ac294244eaa9db5655bc3c3908deb54ac21d4d848caa064ae2774800d360 Dec 05 09:30:25 crc kubenswrapper[4815]: I1205 09:30:25.730245 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:26 crc kubenswrapper[4815]: I1205 09:30:26.560868 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcf9574d-dedc-460f-9799-f556d817fb52","Type":"ContainerStarted","Data":"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6"} Dec 05 09:30:26 crc kubenswrapper[4815]: I1205 09:30:26.561217 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcf9574d-dedc-460f-9799-f556d817fb52","Type":"ContainerStarted","Data":"6616ac294244eaa9db5655bc3c3908deb54ac21d4d848caa064ae2774800d360"} Dec 05 09:30:27 crc kubenswrapper[4815]: I1205 09:30:27.570952 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcf9574d-dedc-460f-9799-f556d817fb52","Type":"ContainerStarted","Data":"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3"} Dec 05 09:30:28 crc kubenswrapper[4815]: I1205 09:30:28.591995 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcf9574d-dedc-460f-9799-f556d817fb52","Type":"ContainerStarted","Data":"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66"} Dec 05 09:30:29 crc kubenswrapper[4815]: I1205 09:30:29.616456 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcf9574d-dedc-460f-9799-f556d817fb52","Type":"ContainerStarted","Data":"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d"} Dec 05 09:30:29 crc kubenswrapper[4815]: I1205 09:30:29.617119 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:30:29 crc kubenswrapper[4815]: I1205 09:30:29.644757 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.479263272 podStartE2EDuration="5.64473165s" podCreationTimestamp="2025-12-05 09:30:24 +0000 UTC" firstStartedPulling="2025-12-05 09:30:25.728030776 +0000 UTC m=+1424.606637603" lastFinishedPulling="2025-12-05 09:30:28.893482754 +0000 UTC m=+1427.772105981" observedRunningTime="2025-12-05 09:30:29.631980591 +0000 UTC m=+1428.510587428" watchObservedRunningTime="2025-12-05 09:30:29.64473165 +0000 UTC m=+1428.523338487" Dec 05 09:30:31 crc kubenswrapper[4815]: I1205 09:30:31.191876 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:31 crc kubenswrapper[4815]: I1205 09:30:31.634880 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="sg-core" containerID="cri-o://e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66" gracePeriod=30 Dec 05 09:30:31 crc kubenswrapper[4815]: I1205 09:30:31.634865 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="ceilometer-central-agent" containerID="cri-o://70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6" gracePeriod=30 Dec 05 09:30:31 crc kubenswrapper[4815]: I1205 09:30:31.634885 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="ceilometer-notification-agent" containerID="cri-o://6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3" gracePeriod=30 Dec 05 09:30:31 crc kubenswrapper[4815]: I1205 09:30:31.634896 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="proxy-httpd" containerID="cri-o://3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d" gracePeriod=30 Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.485826 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.643732 4815 generic.go:334] "Generic (PLEG): container finished" podID="fcf9574d-dedc-460f-9799-f556d817fb52" containerID="3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d" exitCode=0 Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.643766 4815 generic.go:334] "Generic (PLEG): container finished" podID="fcf9574d-dedc-460f-9799-f556d817fb52" containerID="e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66" exitCode=2 Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.643779 4815 generic.go:334] "Generic (PLEG): container finished" podID="fcf9574d-dedc-460f-9799-f556d817fb52" containerID="6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3" exitCode=0 Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.643791 4815 generic.go:334] "Generic (PLEG): container finished" podID="fcf9574d-dedc-460f-9799-f556d817fb52" containerID="70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6" exitCode=0 Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.643833 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcf9574d-dedc-460f-9799-f556d817fb52","Type":"ContainerDied","Data":"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d"} Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.643860 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcf9574d-dedc-460f-9799-f556d817fb52","Type":"ContainerDied","Data":"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66"} Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.643872 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcf9574d-dedc-460f-9799-f556d817fb52","Type":"ContainerDied","Data":"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3"} Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.643883 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcf9574d-dedc-460f-9799-f556d817fb52","Type":"ContainerDied","Data":"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6"} Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.643894 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fcf9574d-dedc-460f-9799-f556d817fb52","Type":"ContainerDied","Data":"6616ac294244eaa9db5655bc3c3908deb54ac21d4d848caa064ae2774800d360"} Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.643913 4815 scope.go:117] "RemoveContainer" containerID="3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.644077 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.664108 4815 scope.go:117] "RemoveContainer" containerID="e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.683198 4815 scope.go:117] "RemoveContainer" containerID="6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.684795 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-run-httpd\") pod \"fcf9574d-dedc-460f-9799-f556d817fb52\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.684934 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-scripts\") pod \"fcf9574d-dedc-460f-9799-f556d817fb52\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.685045 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-sg-core-conf-yaml\") pod \"fcf9574d-dedc-460f-9799-f556d817fb52\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.685140 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-combined-ca-bundle\") pod \"fcf9574d-dedc-460f-9799-f556d817fb52\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.685203 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fcf9574d-dedc-460f-9799-f556d817fb52" (UID: "fcf9574d-dedc-460f-9799-f556d817fb52"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.685221 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-log-httpd\") pod \"fcf9574d-dedc-460f-9799-f556d817fb52\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.685277 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6mkm5\" (UniqueName: \"kubernetes.io/projected/fcf9574d-dedc-460f-9799-f556d817fb52-kube-api-access-6mkm5\") pod \"fcf9574d-dedc-460f-9799-f556d817fb52\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.685307 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-config-data\") pod \"fcf9574d-dedc-460f-9799-f556d817fb52\" (UID: \"fcf9574d-dedc-460f-9799-f556d817fb52\") " Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.685618 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.685759 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fcf9574d-dedc-460f-9799-f556d817fb52" (UID: "fcf9574d-dedc-460f-9799-f556d817fb52"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.691729 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-scripts" (OuterVolumeSpecName: "scripts") pod "fcf9574d-dedc-460f-9799-f556d817fb52" (UID: "fcf9574d-dedc-460f-9799-f556d817fb52"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.694827 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcf9574d-dedc-460f-9799-f556d817fb52-kube-api-access-6mkm5" (OuterVolumeSpecName: "kube-api-access-6mkm5") pod "fcf9574d-dedc-460f-9799-f556d817fb52" (UID: "fcf9574d-dedc-460f-9799-f556d817fb52"). InnerVolumeSpecName "kube-api-access-6mkm5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.705631 4815 scope.go:117] "RemoveContainer" containerID="70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.718680 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fcf9574d-dedc-460f-9799-f556d817fb52" (UID: "fcf9574d-dedc-460f-9799-f556d817fb52"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.751597 4815 scope.go:117] "RemoveContainer" containerID="3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d" Dec 05 09:30:32 crc kubenswrapper[4815]: E1205 09:30:32.759804 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d\": container with ID starting with 3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d not found: ID does not exist" containerID="3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.759867 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d"} err="failed to get container status \"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d\": rpc error: code = NotFound desc = could not find container \"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d\": container with ID starting with 3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.759899 4815 scope.go:117] "RemoveContainer" containerID="e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66" Dec 05 09:30:32 crc kubenswrapper[4815]: E1205 09:30:32.761854 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66\": container with ID starting with e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66 not found: ID does not exist" containerID="e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.761898 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66"} err="failed to get container status \"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66\": rpc error: code = NotFound desc = could not find container \"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66\": container with ID starting with e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.761923 4815 scope.go:117] "RemoveContainer" containerID="6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3" Dec 05 09:30:32 crc kubenswrapper[4815]: E1205 09:30:32.762251 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3\": container with ID starting with 6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3 not found: ID does not exist" containerID="6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.762280 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3"} err="failed to get container status \"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3\": rpc error: code = NotFound desc = could not find container \"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3\": container with ID starting with 6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.762301 4815 scope.go:117] "RemoveContainer" containerID="70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6" Dec 05 09:30:32 crc kubenswrapper[4815]: E1205 09:30:32.762591 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6\": container with ID starting with 70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6 not found: ID does not exist" containerID="70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.762628 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6"} err="failed to get container status \"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6\": rpc error: code = NotFound desc = could not find container \"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6\": container with ID starting with 70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.762647 4815 scope.go:117] "RemoveContainer" containerID="3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.762888 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d"} err="failed to get container status \"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d\": rpc error: code = NotFound desc = could not find container \"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d\": container with ID starting with 3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.762925 4815 scope.go:117] "RemoveContainer" containerID="e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.763544 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66"} err="failed to get container status \"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66\": rpc error: code = NotFound desc = could not find container \"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66\": container with ID starting with e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.763566 4815 scope.go:117] "RemoveContainer" containerID="6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.764155 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3"} err="failed to get container status \"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3\": rpc error: code = NotFound desc = could not find container \"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3\": container with ID starting with 6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.764183 4815 scope.go:117] "RemoveContainer" containerID="70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.765439 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6"} err="failed to get container status \"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6\": rpc error: code = NotFound desc = could not find container \"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6\": container with ID starting with 70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.765467 4815 scope.go:117] "RemoveContainer" containerID="3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.766764 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d"} err="failed to get container status \"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d\": rpc error: code = NotFound desc = could not find container \"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d\": container with ID starting with 3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.766786 4815 scope.go:117] "RemoveContainer" containerID="e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.767594 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66"} err="failed to get container status \"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66\": rpc error: code = NotFound desc = could not find container \"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66\": container with ID starting with e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.767613 4815 scope.go:117] "RemoveContainer" containerID="6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.767841 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3"} err="failed to get container status \"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3\": rpc error: code = NotFound desc = could not find container \"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3\": container with ID starting with 6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.767859 4815 scope.go:117] "RemoveContainer" containerID="70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.768360 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6"} err="failed to get container status \"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6\": rpc error: code = NotFound desc = could not find container \"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6\": container with ID starting with 70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.768449 4815 scope.go:117] "RemoveContainer" containerID="3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.770471 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d"} err="failed to get container status \"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d\": rpc error: code = NotFound desc = could not find container \"3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d\": container with ID starting with 3f7891a353b95873871829ea8446750600cf306b5991d941b63f883ee504ef3d not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.770518 4815 scope.go:117] "RemoveContainer" containerID="e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.771423 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66"} err="failed to get container status \"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66\": rpc error: code = NotFound desc = could not find container \"e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66\": container with ID starting with e238522c6f1643e9dfadbf4de7fef7446d68585d460485055adf3c100c58ea66 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.771481 4815 scope.go:117] "RemoveContainer" containerID="6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.774088 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3"} err="failed to get container status \"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3\": rpc error: code = NotFound desc = could not find container \"6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3\": container with ID starting with 6bb73cd72053f84e93fa88271798fddb9a8a559810147c9317c70cc891c38da3 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.774142 4815 scope.go:117] "RemoveContainer" containerID="70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.774624 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6"} err="failed to get container status \"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6\": rpc error: code = NotFound desc = could not find container \"70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6\": container with ID starting with 70dae3fbabaceb4a2fb4911042c370251a07b535084fb795aabfb257867558f6 not found: ID does not exist" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.787656 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fcf9574d-dedc-460f-9799-f556d817fb52-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.787822 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6mkm5\" (UniqueName: \"kubernetes.io/projected/fcf9574d-dedc-460f-9799-f556d817fb52-kube-api-access-6mkm5\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.787910 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.787993 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.800736 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-config-data" (OuterVolumeSpecName: "config-data") pod "fcf9574d-dedc-460f-9799-f556d817fb52" (UID: "fcf9574d-dedc-460f-9799-f556d817fb52"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.806649 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fcf9574d-dedc-460f-9799-f556d817fb52" (UID: "fcf9574d-dedc-460f-9799-f556d817fb52"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.888660 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.888693 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcf9574d-dedc-460f-9799-f556d817fb52-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:32 crc kubenswrapper[4815]: I1205 09:30:32.999593 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.013559 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.026670 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:33 crc kubenswrapper[4815]: E1205 09:30:33.027094 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="ceilometer-central-agent" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.027112 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="ceilometer-central-agent" Dec 05 09:30:33 crc kubenswrapper[4815]: E1205 09:30:33.027133 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="ceilometer-notification-agent" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.027138 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="ceilometer-notification-agent" Dec 05 09:30:33 crc kubenswrapper[4815]: E1205 09:30:33.027151 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="sg-core" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.027157 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="sg-core" Dec 05 09:30:33 crc kubenswrapper[4815]: E1205 09:30:33.027165 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="proxy-httpd" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.027170 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="proxy-httpd" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.027348 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="proxy-httpd" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.027359 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="ceilometer-central-agent" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.027373 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="sg-core" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.027384 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" containerName="ceilometer-notification-agent" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.028867 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.041556 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.041729 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.050876 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.193961 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.194000 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2swkg\" (UniqueName: \"kubernetes.io/projected/70f070c8-3b5b-4978-972a-c81b38cdac50-kube-api-access-2swkg\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.194044 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-scripts\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.194062 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-run-httpd\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.194085 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-config-data\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.194119 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.194161 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-log-httpd\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.295175 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-log-httpd\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.295244 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.295269 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2swkg\" (UniqueName: \"kubernetes.io/projected/70f070c8-3b5b-4978-972a-c81b38cdac50-kube-api-access-2swkg\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.295315 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-scripts\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.295332 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-run-httpd\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.295358 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-config-data\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.295393 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.295736 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-log-httpd\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.295975 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-run-httpd\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.300050 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.300182 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-scripts\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.302633 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.318330 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-config-data\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.323348 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2swkg\" (UniqueName: \"kubernetes.io/projected/70f070c8-3b5b-4978-972a-c81b38cdac50-kube-api-access-2swkg\") pod \"ceilometer-0\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.372329 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.435945 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcf9574d-dedc-460f-9799-f556d817fb52" path="/var/lib/kubelet/pods/fcf9574d-dedc-460f-9799-f556d817fb52/volumes" Dec 05 09:30:33 crc kubenswrapper[4815]: I1205 09:30:33.952825 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:33 crc kubenswrapper[4815]: W1205 09:30:33.958701 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70f070c8_3b5b_4978_972a_c81b38cdac50.slice/crio-9ded73e3ef8a6b6bf691b49a263eda2a903407f573dc55f9d3b92784c5ccea45 WatchSource:0}: Error finding container 9ded73e3ef8a6b6bf691b49a263eda2a903407f573dc55f9d3b92784c5ccea45: Status 404 returned error can't find the container with id 9ded73e3ef8a6b6bf691b49a263eda2a903407f573dc55f9d3b92784c5ccea45 Dec 05 09:30:34 crc kubenswrapper[4815]: I1205 09:30:34.023686 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:34 crc kubenswrapper[4815]: I1205 09:30:34.689414 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70f070c8-3b5b-4978-972a-c81b38cdac50","Type":"ContainerStarted","Data":"9ded73e3ef8a6b6bf691b49a263eda2a903407f573dc55f9d3b92784c5ccea45"} Dec 05 09:30:34 crc kubenswrapper[4815]: I1205 09:30:34.691610 4815 generic.go:334] "Generic (PLEG): container finished" podID="a9371c4b-130e-4c75-b54a-7dc4d07fe7e4" containerID="ffe9faeaa0bf8cd09661f677d909031160c5256658cc3eb14e3506e73ac4bd46" exitCode=0 Dec 05 09:30:34 crc kubenswrapper[4815]: I1205 09:30:34.691640 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s5qnf" event={"ID":"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4","Type":"ContainerDied","Data":"ffe9faeaa0bf8cd09661f677d909031160c5256658cc3eb14e3506e73ac4bd46"} Dec 05 09:30:35 crc kubenswrapper[4815]: I1205 09:30:35.735594 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70f070c8-3b5b-4978-972a-c81b38cdac50","Type":"ContainerStarted","Data":"421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7"} Dec 05 09:30:35 crc kubenswrapper[4815]: I1205 09:30:35.735912 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70f070c8-3b5b-4978-972a-c81b38cdac50","Type":"ContainerStarted","Data":"2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee"} Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.038725 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.183824 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-config-data\") pod \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.184048 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dfgr\" (UniqueName: \"kubernetes.io/projected/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-kube-api-access-2dfgr\") pod \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.184100 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-scripts\") pod \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.184431 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-combined-ca-bundle\") pod \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\" (UID: \"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4\") " Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.190632 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-scripts" (OuterVolumeSpecName: "scripts") pod "a9371c4b-130e-4c75-b54a-7dc4d07fe7e4" (UID: "a9371c4b-130e-4c75-b54a-7dc4d07fe7e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.192722 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-kube-api-access-2dfgr" (OuterVolumeSpecName: "kube-api-access-2dfgr") pod "a9371c4b-130e-4c75-b54a-7dc4d07fe7e4" (UID: "a9371c4b-130e-4c75-b54a-7dc4d07fe7e4"). InnerVolumeSpecName "kube-api-access-2dfgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.209568 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a9371c4b-130e-4c75-b54a-7dc4d07fe7e4" (UID: "a9371c4b-130e-4c75-b54a-7dc4d07fe7e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.216438 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-config-data" (OuterVolumeSpecName: "config-data") pod "a9371c4b-130e-4c75-b54a-7dc4d07fe7e4" (UID: "a9371c4b-130e-4c75-b54a-7dc4d07fe7e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.286551 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.286600 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.286615 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dfgr\" (UniqueName: \"kubernetes.io/projected/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-kube-api-access-2dfgr\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.286627 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.745311 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-s5qnf" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.745472 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-s5qnf" event={"ID":"a9371c4b-130e-4c75-b54a-7dc4d07fe7e4","Type":"ContainerDied","Data":"ce3f55feafa9f31a392e757f5541eaf6b8f04e9ff15deee6788de3c8c97b0b7d"} Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.746352 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce3f55feafa9f31a392e757f5541eaf6b8f04e9ff15deee6788de3c8c97b0b7d" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.747679 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70f070c8-3b5b-4978-972a-c81b38cdac50","Type":"ContainerStarted","Data":"93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294"} Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.834434 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 09:30:36 crc kubenswrapper[4815]: E1205 09:30:36.834776 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9371c4b-130e-4c75-b54a-7dc4d07fe7e4" containerName="nova-cell0-conductor-db-sync" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.834793 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9371c4b-130e-4c75-b54a-7dc4d07fe7e4" containerName="nova-cell0-conductor-db-sync" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.834955 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9371c4b-130e-4c75-b54a-7dc4d07fe7e4" containerName="nova-cell0-conductor-db-sync" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.835476 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.845101 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-qklpl" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.845101 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.852611 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.995570 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d358cc7-a297-4407-85db-ca619d6dccd4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7d358cc7-a297-4407-85db-ca619d6dccd4\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.995611 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkmwj\" (UniqueName: \"kubernetes.io/projected/7d358cc7-a297-4407-85db-ca619d6dccd4-kube-api-access-jkmwj\") pod \"nova-cell0-conductor-0\" (UID: \"7d358cc7-a297-4407-85db-ca619d6dccd4\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:36 crc kubenswrapper[4815]: I1205 09:30:36.995633 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d358cc7-a297-4407-85db-ca619d6dccd4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7d358cc7-a297-4407-85db-ca619d6dccd4\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.104797 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d358cc7-a297-4407-85db-ca619d6dccd4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7d358cc7-a297-4407-85db-ca619d6dccd4\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.104848 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkmwj\" (UniqueName: \"kubernetes.io/projected/7d358cc7-a297-4407-85db-ca619d6dccd4-kube-api-access-jkmwj\") pod \"nova-cell0-conductor-0\" (UID: \"7d358cc7-a297-4407-85db-ca619d6dccd4\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.104872 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d358cc7-a297-4407-85db-ca619d6dccd4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7d358cc7-a297-4407-85db-ca619d6dccd4\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.108805 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d358cc7-a297-4407-85db-ca619d6dccd4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"7d358cc7-a297-4407-85db-ca619d6dccd4\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.112300 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d358cc7-a297-4407-85db-ca619d6dccd4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"7d358cc7-a297-4407-85db-ca619d6dccd4\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.130146 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkmwj\" (UniqueName: \"kubernetes.io/projected/7d358cc7-a297-4407-85db-ca619d6dccd4-kube-api-access-jkmwj\") pod \"nova-cell0-conductor-0\" (UID: \"7d358cc7-a297-4407-85db-ca619d6dccd4\") " pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.152988 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.606041 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.757472 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70f070c8-3b5b-4978-972a-c81b38cdac50","Type":"ContainerStarted","Data":"375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c"} Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.757919 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="ceilometer-central-agent" containerID="cri-o://2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee" gracePeriod=30 Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.758212 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.758601 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="proxy-httpd" containerID="cri-o://375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c" gracePeriod=30 Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.758658 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="sg-core" containerID="cri-o://93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294" gracePeriod=30 Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.758693 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="ceilometer-notification-agent" containerID="cri-o://421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7" gracePeriod=30 Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.766721 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7d358cc7-a297-4407-85db-ca619d6dccd4","Type":"ContainerStarted","Data":"9c96d92064b47f50cd1411f8c4c3f197accf5a945843e896ea6436f9d443c1f0"} Dec 05 09:30:37 crc kubenswrapper[4815]: I1205 09:30:37.791350 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.701975242 podStartE2EDuration="5.791330429s" podCreationTimestamp="2025-12-05 09:30:32 +0000 UTC" firstStartedPulling="2025-12-05 09:30:33.960456844 +0000 UTC m=+1432.839063681" lastFinishedPulling="2025-12-05 09:30:37.049812031 +0000 UTC m=+1435.928418868" observedRunningTime="2025-12-05 09:30:37.79025992 +0000 UTC m=+1436.668866757" watchObservedRunningTime="2025-12-05 09:30:37.791330429 +0000 UTC m=+1436.669937266" Dec 05 09:30:38 crc kubenswrapper[4815]: I1205 09:30:38.780466 4815 generic.go:334] "Generic (PLEG): container finished" podID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerID="375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c" exitCode=0 Dec 05 09:30:38 crc kubenswrapper[4815]: I1205 09:30:38.780743 4815 generic.go:334] "Generic (PLEG): container finished" podID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerID="93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294" exitCode=2 Dec 05 09:30:38 crc kubenswrapper[4815]: I1205 09:30:38.780753 4815 generic.go:334] "Generic (PLEG): container finished" podID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerID="421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7" exitCode=0 Dec 05 09:30:38 crc kubenswrapper[4815]: I1205 09:30:38.780794 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70f070c8-3b5b-4978-972a-c81b38cdac50","Type":"ContainerDied","Data":"375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c"} Dec 05 09:30:38 crc kubenswrapper[4815]: I1205 09:30:38.780819 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70f070c8-3b5b-4978-972a-c81b38cdac50","Type":"ContainerDied","Data":"93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294"} Dec 05 09:30:38 crc kubenswrapper[4815]: I1205 09:30:38.780828 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70f070c8-3b5b-4978-972a-c81b38cdac50","Type":"ContainerDied","Data":"421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7"} Dec 05 09:30:38 crc kubenswrapper[4815]: I1205 09:30:38.782349 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"7d358cc7-a297-4407-85db-ca619d6dccd4","Type":"ContainerStarted","Data":"b42a07766991cc80dc55a926ae1da2e01e5a47252d40db8aa9f06945556915ac"} Dec 05 09:30:38 crc kubenswrapper[4815]: I1205 09:30:38.783059 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:38 crc kubenswrapper[4815]: I1205 09:30:38.806248 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.806196618 podStartE2EDuration="2.806196618s" podCreationTimestamp="2025-12-05 09:30:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:30:38.799948703 +0000 UTC m=+1437.678555540" watchObservedRunningTime="2025-12-05 09:30:38.806196618 +0000 UTC m=+1437.684803445" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.718561 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.810844 4815 generic.go:334] "Generic (PLEG): container finished" podID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerID="2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee" exitCode=0 Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.810884 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70f070c8-3b5b-4978-972a-c81b38cdac50","Type":"ContainerDied","Data":"2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee"} Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.810921 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"70f070c8-3b5b-4978-972a-c81b38cdac50","Type":"ContainerDied","Data":"9ded73e3ef8a6b6bf691b49a263eda2a903407f573dc55f9d3b92784c5ccea45"} Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.810938 4815 scope.go:117] "RemoveContainer" containerID="375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.811201 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.848765 4815 scope.go:117] "RemoveContainer" containerID="93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.872059 4815 scope.go:117] "RemoveContainer" containerID="421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.889786 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-config-data\") pod \"70f070c8-3b5b-4978-972a-c81b38cdac50\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.889854 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-run-httpd\") pod \"70f070c8-3b5b-4978-972a-c81b38cdac50\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.889919 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-combined-ca-bundle\") pod \"70f070c8-3b5b-4978-972a-c81b38cdac50\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.890001 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-sg-core-conf-yaml\") pod \"70f070c8-3b5b-4978-972a-c81b38cdac50\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.890081 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-log-httpd\") pod \"70f070c8-3b5b-4978-972a-c81b38cdac50\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.890158 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2swkg\" (UniqueName: \"kubernetes.io/projected/70f070c8-3b5b-4978-972a-c81b38cdac50-kube-api-access-2swkg\") pod \"70f070c8-3b5b-4978-972a-c81b38cdac50\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.890223 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-scripts\") pod \"70f070c8-3b5b-4978-972a-c81b38cdac50\" (UID: \"70f070c8-3b5b-4978-972a-c81b38cdac50\") " Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.891118 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "70f070c8-3b5b-4978-972a-c81b38cdac50" (UID: "70f070c8-3b5b-4978-972a-c81b38cdac50"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.891969 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "70f070c8-3b5b-4978-972a-c81b38cdac50" (UID: "70f070c8-3b5b-4978-972a-c81b38cdac50"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.897764 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70f070c8-3b5b-4978-972a-c81b38cdac50-kube-api-access-2swkg" (OuterVolumeSpecName: "kube-api-access-2swkg") pod "70f070c8-3b5b-4978-972a-c81b38cdac50" (UID: "70f070c8-3b5b-4978-972a-c81b38cdac50"). InnerVolumeSpecName "kube-api-access-2swkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.908165 4815 scope.go:117] "RemoveContainer" containerID="2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.913223 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-scripts" (OuterVolumeSpecName: "scripts") pod "70f070c8-3b5b-4978-972a-c81b38cdac50" (UID: "70f070c8-3b5b-4978-972a-c81b38cdac50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.916013 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "70f070c8-3b5b-4978-972a-c81b38cdac50" (UID: "70f070c8-3b5b-4978-972a-c81b38cdac50"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.962444 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "70f070c8-3b5b-4978-972a-c81b38cdac50" (UID: "70f070c8-3b5b-4978-972a-c81b38cdac50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.972959 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-config-data" (OuterVolumeSpecName: "config-data") pod "70f070c8-3b5b-4978-972a-c81b38cdac50" (UID: "70f070c8-3b5b-4978-972a-c81b38cdac50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.993748 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.993791 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.993806 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2swkg\" (UniqueName: \"kubernetes.io/projected/70f070c8-3b5b-4978-972a-c81b38cdac50-kube-api-access-2swkg\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.993819 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.993831 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.993841 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/70f070c8-3b5b-4978-972a-c81b38cdac50-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.993851 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70f070c8-3b5b-4978-972a-c81b38cdac50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.998074 4815 scope.go:117] "RemoveContainer" containerID="375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c" Dec 05 09:30:41 crc kubenswrapper[4815]: E1205 09:30:41.998551 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c\": container with ID starting with 375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c not found: ID does not exist" containerID="375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.998584 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c"} err="failed to get container status \"375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c\": rpc error: code = NotFound desc = could not find container \"375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c\": container with ID starting with 375f7b3b56c449e40e9c89ddd189e7fc869799d09e32d92c4eb431d4a862b75c not found: ID does not exist" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.998608 4815 scope.go:117] "RemoveContainer" containerID="93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294" Dec 05 09:30:41 crc kubenswrapper[4815]: E1205 09:30:41.998913 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294\": container with ID starting with 93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294 not found: ID does not exist" containerID="93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.998933 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294"} err="failed to get container status \"93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294\": rpc error: code = NotFound desc = could not find container \"93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294\": container with ID starting with 93026737039e3706b8255150210d588dde5a4e7c547a95dba68ca0537f0c6294 not found: ID does not exist" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.998946 4815 scope.go:117] "RemoveContainer" containerID="421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7" Dec 05 09:30:41 crc kubenswrapper[4815]: E1205 09:30:41.999214 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7\": container with ID starting with 421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7 not found: ID does not exist" containerID="421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.999257 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7"} err="failed to get container status \"421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7\": rpc error: code = NotFound desc = could not find container \"421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7\": container with ID starting with 421414b3f3038cb889cfb5d1b076abe4d3ef43e8265305a758208464a5d369e7 not found: ID does not exist" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.999275 4815 scope.go:117] "RemoveContainer" containerID="2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee" Dec 05 09:30:41 crc kubenswrapper[4815]: E1205 09:30:41.999563 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee\": container with ID starting with 2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee not found: ID does not exist" containerID="2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee" Dec 05 09:30:41 crc kubenswrapper[4815]: I1205 09:30:41.999591 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee"} err="failed to get container status \"2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee\": rpc error: code = NotFound desc = could not find container \"2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee\": container with ID starting with 2c975124eea4f8e21c5e89ab053470321aea0c9a3803beb86d44ff3c0bfbb2ee not found: ID does not exist" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.170385 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.220156 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.220251 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.232059 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:42 crc kubenswrapper[4815]: E1205 09:30:42.232480 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="ceilometer-notification-agent" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.232512 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="ceilometer-notification-agent" Dec 05 09:30:42 crc kubenswrapper[4815]: E1205 09:30:42.232526 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="ceilometer-central-agent" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.232532 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="ceilometer-central-agent" Dec 05 09:30:42 crc kubenswrapper[4815]: E1205 09:30:42.232549 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="proxy-httpd" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.232555 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="proxy-httpd" Dec 05 09:30:42 crc kubenswrapper[4815]: E1205 09:30:42.232574 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="sg-core" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.232580 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="sg-core" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.232773 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="ceilometer-central-agent" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.232788 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="ceilometer-notification-agent" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.232795 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="sg-core" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.232814 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" containerName="proxy-httpd" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.234338 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.238479 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.239047 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.245434 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.421289 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-scripts\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.421362 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.421414 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-config-data\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.421448 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-run-httpd\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.421474 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9hkb\" (UniqueName: \"kubernetes.io/projected/5cf6b21f-8ae2-49db-a48c-1fc941793393-kube-api-access-x9hkb\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.421538 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-log-httpd\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.421584 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.523053 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-scripts\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.523141 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.523176 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-config-data\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.523196 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-run-httpd\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.523215 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9hkb\" (UniqueName: \"kubernetes.io/projected/5cf6b21f-8ae2-49db-a48c-1fc941793393-kube-api-access-x9hkb\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.523272 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-log-httpd\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.523341 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.524856 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-log-httpd\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.525232 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-run-httpd\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.528428 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.528619 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.531011 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-config-data\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.543811 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-scripts\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.546603 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9hkb\" (UniqueName: \"kubernetes.io/projected/5cf6b21f-8ae2-49db-a48c-1fc941793393-kube-api-access-x9hkb\") pod \"ceilometer-0\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.562419 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.700014 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-jxrzv"] Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.701360 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.704300 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.721771 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.724188 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jxrzv"] Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.830203 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-config-data\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.830369 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-scripts\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.830401 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.830523 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlxm8\" (UniqueName: \"kubernetes.io/projected/04a96446-9720-4d97-a657-42ab3826aee3-kube-api-access-jlxm8\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.931890 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlxm8\" (UniqueName: \"kubernetes.io/projected/04a96446-9720-4d97-a657-42ab3826aee3-kube-api-access-jlxm8\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.931951 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-config-data\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.932033 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-scripts\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.932053 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.948221 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-scripts\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.948895 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-config-data\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.961617 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:42 crc kubenswrapper[4815]: I1205 09:30:42.990200 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlxm8\" (UniqueName: \"kubernetes.io/projected/04a96446-9720-4d97-a657-42ab3826aee3-kube-api-access-jlxm8\") pod \"nova-cell0-cell-mapping-jxrzv\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.004033 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.005595 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.016619 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.047041 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.067300 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.101013 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.102142 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.111877 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.117897 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.146153 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90293839-ab90-42b5-ba19-880e4fa44624-logs\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.146418 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.146713 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djd6w\" (UniqueName: \"kubernetes.io/projected/90293839-ab90-42b5-ba19-880e4fa44624-kube-api-access-djd6w\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.146835 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-config-data\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.211539 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.213097 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.218972 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.248810 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-config-data\") pod \"nova-scheduler-0\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " pod="openstack/nova-scheduler-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.248897 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90293839-ab90-42b5-ba19-880e4fa44624-logs\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.248939 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.248977 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlxkg\" (UniqueName: \"kubernetes.io/projected/58342a1e-cd29-49e1-b0f5-815a6508c7de-kube-api-access-dlxkg\") pod \"nova-scheduler-0\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " pod="openstack/nova-scheduler-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.249030 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " pod="openstack/nova-scheduler-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.249102 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djd6w\" (UniqueName: \"kubernetes.io/projected/90293839-ab90-42b5-ba19-880e4fa44624-kube-api-access-djd6w\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.249152 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-config-data\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.254119 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.254382 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90293839-ab90-42b5-ba19-880e4fa44624-logs\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.269341 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-config-data\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.342093 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.353518 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " pod="openstack/nova-scheduler-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.353925 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtmkg\" (UniqueName: \"kubernetes.io/projected/53aabb1e-7201-486d-96ca-076bd6cc7528-kube-api-access-mtmkg\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.354035 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-config-data\") pod \"nova-scheduler-0\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " pod="openstack/nova-scheduler-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.354153 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53aabb1e-7201-486d-96ca-076bd6cc7528-logs\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.354268 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlxkg\" (UniqueName: \"kubernetes.io/projected/58342a1e-cd29-49e1-b0f5-815a6508c7de-kube-api-access-dlxkg\") pod \"nova-scheduler-0\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " pod="openstack/nova-scheduler-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.354364 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.354481 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-config-data\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.358299 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " pod="openstack/nova-scheduler-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.367157 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-config-data\") pod \"nova-scheduler-0\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " pod="openstack/nova-scheduler-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.387577 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.454344 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlxkg\" (UniqueName: \"kubernetes.io/projected/58342a1e-cd29-49e1-b0f5-815a6508c7de-kube-api-access-dlxkg\") pod \"nova-scheduler-0\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " pod="openstack/nova-scheduler-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.460244 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53aabb1e-7201-486d-96ca-076bd6cc7528-logs\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.474878 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.475042 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-config-data\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.475441 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtmkg\" (UniqueName: \"kubernetes.io/projected/53aabb1e-7201-486d-96ca-076bd6cc7528-kube-api-access-mtmkg\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.466400 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53aabb1e-7201-486d-96ca-076bd6cc7528-logs\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.493543 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-config-data\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.495303 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.537954 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.625319 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70f070c8-3b5b-4978-972a-c81b38cdac50" path="/var/lib/kubelet/pods/70f070c8-3b5b-4978-972a-c81b38cdac50/volumes" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.629556 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djd6w\" (UniqueName: \"kubernetes.io/projected/90293839-ab90-42b5-ba19-880e4fa44624-kube-api-access-djd6w\") pod \"nova-api-0\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.636259 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtmkg\" (UniqueName: \"kubernetes.io/projected/53aabb1e-7201-486d-96ca-076bd6cc7528-kube-api-access-mtmkg\") pod \"nova-metadata-0\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.647853 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-f98bz"] Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.649469 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.701889 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-f98bz"] Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.728724 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.730044 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.731147 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.738288 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.769656 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.826923 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-nb\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.827057 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-dns-svc\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.827090 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.827111 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88ghq\" (UniqueName: \"kubernetes.io/projected/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-kube-api-access-88ghq\") pod \"nova-cell1-novncproxy-0\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.827302 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.827351 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-config\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.827382 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-sb\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.827420 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8k4h\" (UniqueName: \"kubernetes.io/projected/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-kube-api-access-l8k4h\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.848932 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.933987 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.934362 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-config\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.934404 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-sb\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.934442 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8k4h\" (UniqueName: \"kubernetes.io/projected/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-kube-api-access-l8k4h\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.934524 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-nb\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.934582 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-dns-svc\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.934608 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.934635 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88ghq\" (UniqueName: \"kubernetes.io/projected/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-kube-api-access-88ghq\") pod \"nova-cell1-novncproxy-0\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.935319 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-config\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.935644 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-nb\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.936910 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-dns-svc\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.936974 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-sb\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.945670 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cf6b21f-8ae2-49db-a48c-1fc941793393","Type":"ContainerStarted","Data":"e6efc06cd0731cc93373e5ad7bf6870a5978561ba95adf650c0b27eb2907b16d"} Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.950273 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.952848 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:43 crc kubenswrapper[4815]: I1205 09:30:43.986370 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88ghq\" (UniqueName: \"kubernetes.io/projected/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-kube-api-access-88ghq\") pod \"nova-cell1-novncproxy-0\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:44 crc kubenswrapper[4815]: I1205 09:30:44.013399 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8k4h\" (UniqueName: \"kubernetes.io/projected/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-kube-api-access-l8k4h\") pod \"dnsmasq-dns-55cbc7dc9c-f98bz\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:44 crc kubenswrapper[4815]: I1205 09:30:44.042221 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:44 crc kubenswrapper[4815]: I1205 09:30:44.137040 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:44 crc kubenswrapper[4815]: I1205 09:30:44.359953 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-jxrzv"] Dec 05 09:30:44 crc kubenswrapper[4815]: I1205 09:30:44.577277 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:30:44 crc kubenswrapper[4815]: I1205 09:30:44.689445 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:30:44 crc kubenswrapper[4815]: I1205 09:30:44.816692 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:30:44 crc kubenswrapper[4815]: I1205 09:30:44.828752 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-f98bz"] Dec 05 09:30:44 crc kubenswrapper[4815]: I1205 09:30:44.981826 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"53aabb1e-7201-486d-96ca-076bd6cc7528","Type":"ContainerStarted","Data":"a15462848d1bb8f112a115a2a5427db31aaf2a5d46f866ea0663e88b63908d0f"} Dec 05 09:30:44 crc kubenswrapper[4815]: I1205 09:30:44.998041 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90293839-ab90-42b5-ba19-880e4fa44624","Type":"ContainerStarted","Data":"8b160291575a4af50952ffe2dde448295cbc76ccfeb9e976b2489dc454c552ea"} Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.001712 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jxrzv" event={"ID":"04a96446-9720-4d97-a657-42ab3826aee3","Type":"ContainerStarted","Data":"f5127d3ec099ece6f2dc2b9875ac7d19cfe4af3a2626d723607dcf76d07b7498"} Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.001755 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jxrzv" event={"ID":"04a96446-9720-4d97-a657-42ab3826aee3","Type":"ContainerStarted","Data":"fd6f20585712ca74858f7ae45e524f0cfba5a1c41e5bec48c250ec1096f892bc"} Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.015971 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cf6b21f-8ae2-49db-a48c-1fc941793393","Type":"ContainerStarted","Data":"59804dc53de4af87ec676cfdf4914a1b245b6034588ffc6f3f06542908860e2b"} Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.036270 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58342a1e-cd29-49e1-b0f5-815a6508c7de","Type":"ContainerStarted","Data":"d003a2068447d5e5f7181471d830e73bca83b86886416b5edac8cfe97150e4ff"} Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.044162 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" event={"ID":"7584b3a2-aaea-4f2e-b694-269eaa08bbc2","Type":"ContainerStarted","Data":"33cc355a535d54973ccf741a2c6eaeddb51b5378e284fd9126f2d69204902c0b"} Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.105034 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-jxrzv" podStartSLOduration=3.105017782 podStartE2EDuration="3.105017782s" podCreationTimestamp="2025-12-05 09:30:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:30:45.03340298 +0000 UTC m=+1443.912009837" watchObservedRunningTime="2025-12-05 09:30:45.105017782 +0000 UTC m=+1443.983624619" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.118001 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.216369 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dxtfw"] Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.217482 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.220178 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.220217 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.226401 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dxtfw"] Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.319318 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95zds\" (UniqueName: \"kubernetes.io/projected/6f42cc49-79a7-494e-9f3e-da2da9da9940-kube-api-access-95zds\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.319365 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.319433 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-scripts\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.319518 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-config-data\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.420721 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-config-data\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.420880 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95zds\" (UniqueName: \"kubernetes.io/projected/6f42cc49-79a7-494e-9f3e-da2da9da9940-kube-api-access-95zds\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.420913 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.420980 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-scripts\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.427441 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-scripts\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.427797 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.436863 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-config-data\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.454814 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95zds\" (UniqueName: \"kubernetes.io/projected/6f42cc49-79a7-494e-9f3e-da2da9da9940-kube-api-access-95zds\") pod \"nova-cell1-conductor-db-sync-dxtfw\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:45 crc kubenswrapper[4815]: I1205 09:30:45.548599 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:46 crc kubenswrapper[4815]: I1205 09:30:46.040842 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dxtfw"] Dec 05 09:30:46 crc kubenswrapper[4815]: I1205 09:30:46.101009 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3cdb4e0e-b110-4319-b2e0-c20ef14f0525","Type":"ContainerStarted","Data":"610e760da897bdea1cda35a7f0587c945a55b45695dced1df510ea6f097f7e79"} Dec 05 09:30:46 crc kubenswrapper[4815]: I1205 09:30:46.121730 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cf6b21f-8ae2-49db-a48c-1fc941793393","Type":"ContainerStarted","Data":"36b84d577da0326ee1081b671f6786c7c5e164a030fed329d1429bfd9260ff21"} Dec 05 09:30:46 crc kubenswrapper[4815]: I1205 09:30:46.127550 4815 generic.go:334] "Generic (PLEG): container finished" podID="7584b3a2-aaea-4f2e-b694-269eaa08bbc2" containerID="b2ea99432ce086b24bb1c63de777a21b8b9837b7e8d7ac6d2082b9ed512fb66a" exitCode=0 Dec 05 09:30:46 crc kubenswrapper[4815]: I1205 09:30:46.127653 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" event={"ID":"7584b3a2-aaea-4f2e-b694-269eaa08bbc2","Type":"ContainerDied","Data":"b2ea99432ce086b24bb1c63de777a21b8b9837b7e8d7ac6d2082b9ed512fb66a"} Dec 05 09:30:47 crc kubenswrapper[4815]: I1205 09:30:47.147709 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dxtfw" event={"ID":"6f42cc49-79a7-494e-9f3e-da2da9da9940","Type":"ContainerStarted","Data":"41ea05b3ae7a09bbe7565edab94ed72b396859a198682ebbcb08d99b0730e0ba"} Dec 05 09:30:47 crc kubenswrapper[4815]: I1205 09:30:47.148088 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dxtfw" event={"ID":"6f42cc49-79a7-494e-9f3e-da2da9da9940","Type":"ContainerStarted","Data":"532361772e59eceac99147ea35ea6fb7eefff2e6443eb87a3f23fd887a199cfb"} Dec 05 09:30:47 crc kubenswrapper[4815]: I1205 09:30:47.155683 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cf6b21f-8ae2-49db-a48c-1fc941793393","Type":"ContainerStarted","Data":"2b5f013e7927829e18e66a813af08f73744307f0cb29bf8bf981419c6756125a"} Dec 05 09:30:47 crc kubenswrapper[4815]: I1205 09:30:47.169717 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" event={"ID":"7584b3a2-aaea-4f2e-b694-269eaa08bbc2","Type":"ContainerStarted","Data":"2129ec2bb316370947a664c7161bd798f9a3d150d45ccdad272ff5342c505d09"} Dec 05 09:30:47 crc kubenswrapper[4815]: I1205 09:30:47.171094 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:47 crc kubenswrapper[4815]: I1205 09:30:47.174080 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-dxtfw" podStartSLOduration=2.174062705 podStartE2EDuration="2.174062705s" podCreationTimestamp="2025-12-05 09:30:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:30:47.166713389 +0000 UTC m=+1446.045320256" watchObservedRunningTime="2025-12-05 09:30:47.174062705 +0000 UTC m=+1446.052669542" Dec 05 09:30:47 crc kubenswrapper[4815]: I1205 09:30:47.200213 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" podStartSLOduration=4.200192679 podStartE2EDuration="4.200192679s" podCreationTimestamp="2025-12-05 09:30:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:30:47.198766151 +0000 UTC m=+1446.077372988" watchObservedRunningTime="2025-12-05 09:30:47.200192679 +0000 UTC m=+1446.078799516" Dec 05 09:30:47 crc kubenswrapper[4815]: I1205 09:30:47.417354 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:30:47 crc kubenswrapper[4815]: I1205 09:30:47.470846 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.217362 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3cdb4e0e-b110-4319-b2e0-c20ef14f0525","Type":"ContainerStarted","Data":"ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a"} Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.217437 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="3cdb4e0e-b110-4319-b2e0-c20ef14f0525" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a" gracePeriod=30 Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.221735 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cf6b21f-8ae2-49db-a48c-1fc941793393","Type":"ContainerStarted","Data":"da11f3e7790c67a5bda952d7353f8a158de3061914204669a538fd291842c7d9"} Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.222615 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.225870 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58342a1e-cd29-49e1-b0f5-815a6508c7de","Type":"ContainerStarted","Data":"450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376"} Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.228573 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"53aabb1e-7201-486d-96ca-076bd6cc7528","Type":"ContainerStarted","Data":"d5c0c3760c68525922d428fcdbaaf27008427f6e305e7c1fd79a88ecfc6bcb6e"} Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.228613 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"53aabb1e-7201-486d-96ca-076bd6cc7528","Type":"ContainerStarted","Data":"50ac9f681b3402aa72ac2afb224cc1d323d353143ae3756f6c2e256f5d14f62f"} Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.228652 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="53aabb1e-7201-486d-96ca-076bd6cc7528" containerName="nova-metadata-log" containerID="cri-o://50ac9f681b3402aa72ac2afb224cc1d323d353143ae3756f6c2e256f5d14f62f" gracePeriod=30 Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.228670 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="53aabb1e-7201-486d-96ca-076bd6cc7528" containerName="nova-metadata-metadata" containerID="cri-o://d5c0c3760c68525922d428fcdbaaf27008427f6e305e7c1fd79a88ecfc6bcb6e" gracePeriod=30 Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.232961 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90293839-ab90-42b5-ba19-880e4fa44624","Type":"ContainerStarted","Data":"631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42"} Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.232998 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90293839-ab90-42b5-ba19-880e4fa44624","Type":"ContainerStarted","Data":"cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d"} Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.245586 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.4356018280000002 podStartE2EDuration="8.245571402s" podCreationTimestamp="2025-12-05 09:30:43 +0000 UTC" firstStartedPulling="2025-12-05 09:30:45.135733417 +0000 UTC m=+1444.014340254" lastFinishedPulling="2025-12-05 09:30:49.945702981 +0000 UTC m=+1448.824309828" observedRunningTime="2025-12-05 09:30:51.236686486 +0000 UTC m=+1450.115293323" watchObservedRunningTime="2025-12-05 09:30:51.245571402 +0000 UTC m=+1450.124178239" Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.266642 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.135183108 podStartE2EDuration="8.266623281s" podCreationTimestamp="2025-12-05 09:30:43 +0000 UTC" firstStartedPulling="2025-12-05 09:30:44.822334163 +0000 UTC m=+1443.700941000" lastFinishedPulling="2025-12-05 09:30:49.953774316 +0000 UTC m=+1448.832381173" observedRunningTime="2025-12-05 09:30:51.260120649 +0000 UTC m=+1450.138727486" watchObservedRunningTime="2025-12-05 09:30:51.266623281 +0000 UTC m=+1450.145230118" Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.280871 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.935611346 podStartE2EDuration="8.280854829s" podCreationTimestamp="2025-12-05 09:30:43 +0000 UTC" firstStartedPulling="2025-12-05 09:30:44.607271 +0000 UTC m=+1443.485877837" lastFinishedPulling="2025-12-05 09:30:49.952514483 +0000 UTC m=+1448.831121320" observedRunningTime="2025-12-05 09:30:51.279873544 +0000 UTC m=+1450.158480381" watchObservedRunningTime="2025-12-05 09:30:51.280854829 +0000 UTC m=+1450.159461666" Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.313812 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.706378493 podStartE2EDuration="9.313789744s" podCreationTimestamp="2025-12-05 09:30:42 +0000 UTC" firstStartedPulling="2025-12-05 09:30:43.314746875 +0000 UTC m=+1442.193353712" lastFinishedPulling="2025-12-05 09:30:49.922158106 +0000 UTC m=+1448.800764963" observedRunningTime="2025-12-05 09:30:51.30385656 +0000 UTC m=+1450.182463397" watchObservedRunningTime="2025-12-05 09:30:51.313789744 +0000 UTC m=+1450.192396581" Dec 05 09:30:51 crc kubenswrapper[4815]: I1205 09:30:51.363418 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.140870359 podStartE2EDuration="9.363398293s" podCreationTimestamp="2025-12-05 09:30:42 +0000 UTC" firstStartedPulling="2025-12-05 09:30:44.721107433 +0000 UTC m=+1443.599714270" lastFinishedPulling="2025-12-05 09:30:49.943635357 +0000 UTC m=+1448.822242204" observedRunningTime="2025-12-05 09:30:51.341389478 +0000 UTC m=+1450.219996335" watchObservedRunningTime="2025-12-05 09:30:51.363398293 +0000 UTC m=+1450.242005130" Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.243927 4815 generic.go:334] "Generic (PLEG): container finished" podID="53aabb1e-7201-486d-96ca-076bd6cc7528" containerID="d5c0c3760c68525922d428fcdbaaf27008427f6e305e7c1fd79a88ecfc6bcb6e" exitCode=0 Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.244176 4815 generic.go:334] "Generic (PLEG): container finished" podID="53aabb1e-7201-486d-96ca-076bd6cc7528" containerID="50ac9f681b3402aa72ac2afb224cc1d323d353143ae3756f6c2e256f5d14f62f" exitCode=143 Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.244139 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"53aabb1e-7201-486d-96ca-076bd6cc7528","Type":"ContainerDied","Data":"d5c0c3760c68525922d428fcdbaaf27008427f6e305e7c1fd79a88ecfc6bcb6e"} Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.245154 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"53aabb1e-7201-486d-96ca-076bd6cc7528","Type":"ContainerDied","Data":"50ac9f681b3402aa72ac2afb224cc1d323d353143ae3756f6c2e256f5d14f62f"} Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.420864 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.503907 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-config-data\") pod \"53aabb1e-7201-486d-96ca-076bd6cc7528\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.504788 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-combined-ca-bundle\") pod \"53aabb1e-7201-486d-96ca-076bd6cc7528\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.505100 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53aabb1e-7201-486d-96ca-076bd6cc7528-logs\") pod \"53aabb1e-7201-486d-96ca-076bd6cc7528\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.505558 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/53aabb1e-7201-486d-96ca-076bd6cc7528-logs" (OuterVolumeSpecName: "logs") pod "53aabb1e-7201-486d-96ca-076bd6cc7528" (UID: "53aabb1e-7201-486d-96ca-076bd6cc7528"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.505993 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtmkg\" (UniqueName: \"kubernetes.io/projected/53aabb1e-7201-486d-96ca-076bd6cc7528-kube-api-access-mtmkg\") pod \"53aabb1e-7201-486d-96ca-076bd6cc7528\" (UID: \"53aabb1e-7201-486d-96ca-076bd6cc7528\") " Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.506981 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53aabb1e-7201-486d-96ca-076bd6cc7528-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.517738 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/53aabb1e-7201-486d-96ca-076bd6cc7528-kube-api-access-mtmkg" (OuterVolumeSpecName: "kube-api-access-mtmkg") pod "53aabb1e-7201-486d-96ca-076bd6cc7528" (UID: "53aabb1e-7201-486d-96ca-076bd6cc7528"). InnerVolumeSpecName "kube-api-access-mtmkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.537631 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-config-data" (OuterVolumeSpecName: "config-data") pod "53aabb1e-7201-486d-96ca-076bd6cc7528" (UID: "53aabb1e-7201-486d-96ca-076bd6cc7528"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.579932 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "53aabb1e-7201-486d-96ca-076bd6cc7528" (UID: "53aabb1e-7201-486d-96ca-076bd6cc7528"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.609005 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.609054 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtmkg\" (UniqueName: \"kubernetes.io/projected/53aabb1e-7201-486d-96ca-076bd6cc7528-kube-api-access-mtmkg\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:52 crc kubenswrapper[4815]: I1205 09:30:52.609066 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53aabb1e-7201-486d-96ca-076bd6cc7528-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.254028 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.254032 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"53aabb1e-7201-486d-96ca-076bd6cc7528","Type":"ContainerDied","Data":"a15462848d1bb8f112a115a2a5427db31aaf2a5d46f866ea0663e88b63908d0f"} Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.255073 4815 scope.go:117] "RemoveContainer" containerID="d5c0c3760c68525922d428fcdbaaf27008427f6e305e7c1fd79a88ecfc6bcb6e" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.282304 4815 scope.go:117] "RemoveContainer" containerID="50ac9f681b3402aa72ac2afb224cc1d323d353143ae3756f6c2e256f5d14f62f" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.310763 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.321342 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.334364 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:30:53 crc kubenswrapper[4815]: E1205 09:30:53.335057 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53aabb1e-7201-486d-96ca-076bd6cc7528" containerName="nova-metadata-log" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.335155 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="53aabb1e-7201-486d-96ca-076bd6cc7528" containerName="nova-metadata-log" Dec 05 09:30:53 crc kubenswrapper[4815]: E1205 09:30:53.335274 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="53aabb1e-7201-486d-96ca-076bd6cc7528" containerName="nova-metadata-metadata" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.335355 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="53aabb1e-7201-486d-96ca-076bd6cc7528" containerName="nova-metadata-metadata" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.335664 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="53aabb1e-7201-486d-96ca-076bd6cc7528" containerName="nova-metadata-metadata" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.335776 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="53aabb1e-7201-486d-96ca-076bd6cc7528" containerName="nova-metadata-log" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.336920 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.340449 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.340740 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.350782 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.421667 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-config-data\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.421714 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.421816 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf747dae-9ee6-42c5-a93d-a944f364c073-logs\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.421866 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.421924 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4tmb\" (UniqueName: \"kubernetes.io/projected/cf747dae-9ee6-42c5-a93d-a944f364c073-kube-api-access-b4tmb\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.429506 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="53aabb1e-7201-486d-96ca-076bd6cc7528" path="/var/lib/kubelet/pods/53aabb1e-7201-486d-96ca-076bd6cc7528/volumes" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.505323 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.506472 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.522403 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4tmb\" (UniqueName: \"kubernetes.io/projected/cf747dae-9ee6-42c5-a93d-a944f364c073-kube-api-access-b4tmb\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.522724 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-config-data\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.522823 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.523099 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf747dae-9ee6-42c5-a93d-a944f364c073-logs\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.523236 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.533568 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf747dae-9ee6-42c5-a93d-a944f364c073-logs\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.552231 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-config-data\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.555046 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.555753 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.569202 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4tmb\" (UniqueName: \"kubernetes.io/projected/cf747dae-9ee6-42c5-a93d-a944f364c073-kube-api-access-b4tmb\") pod \"nova-metadata-0\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.689977 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.705291 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.730046 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 09:30:53 crc kubenswrapper[4815]: I1205 09:30:53.730287 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 09:30:54 crc kubenswrapper[4815]: I1205 09:30:54.045701 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:30:54 crc kubenswrapper[4815]: I1205 09:30:54.126340 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-799db77f-npk6z"] Dec 05 09:30:54 crc kubenswrapper[4815]: I1205 09:30:54.126616 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-799db77f-npk6z" podUID="f042ee2e-f5c8-46af-b494-523064e9b7cd" containerName="dnsmasq-dns" containerID="cri-o://7f2eb5874d49f85f7ac50f51efe41dfd444da805af11a9738249a868e55988a6" gracePeriod=10 Dec 05 09:30:54 crc kubenswrapper[4815]: I1205 09:30:54.139594 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:30:54 crc kubenswrapper[4815]: I1205 09:30:54.295296 4815 generic.go:334] "Generic (PLEG): container finished" podID="f042ee2e-f5c8-46af-b494-523064e9b7cd" containerID="7f2eb5874d49f85f7ac50f51efe41dfd444da805af11a9738249a868e55988a6" exitCode=0 Dec 05 09:30:54 crc kubenswrapper[4815]: I1205 09:30:54.296365 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-799db77f-npk6z" event={"ID":"f042ee2e-f5c8-46af-b494-523064e9b7cd","Type":"ContainerDied","Data":"7f2eb5874d49f85f7ac50f51efe41dfd444da805af11a9738249a868e55988a6"} Dec 05 09:30:54 crc kubenswrapper[4815]: I1205 09:30:54.326769 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:30:54 crc kubenswrapper[4815]: I1205 09:30:54.338645 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 09:30:54 crc kubenswrapper[4815]: I1205 09:30:54.814642 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="90293839-ab90-42b5-ba19-880e4fa44624" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.168:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:30:54 crc kubenswrapper[4815]: I1205 09:30:54.814955 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="90293839-ab90-42b5-ba19-880e4fa44624" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.168:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.048899 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.137861 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27nn4\" (UniqueName: \"kubernetes.io/projected/f042ee2e-f5c8-46af-b494-523064e9b7cd-kube-api-access-27nn4\") pod \"f042ee2e-f5c8-46af-b494-523064e9b7cd\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.137920 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-config\") pod \"f042ee2e-f5c8-46af-b494-523064e9b7cd\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.137950 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-nb\") pod \"f042ee2e-f5c8-46af-b494-523064e9b7cd\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.138047 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-sb\") pod \"f042ee2e-f5c8-46af-b494-523064e9b7cd\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.138092 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-dns-svc\") pod \"f042ee2e-f5c8-46af-b494-523064e9b7cd\" (UID: \"f042ee2e-f5c8-46af-b494-523064e9b7cd\") " Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.153305 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f042ee2e-f5c8-46af-b494-523064e9b7cd-kube-api-access-27nn4" (OuterVolumeSpecName: "kube-api-access-27nn4") pod "f042ee2e-f5c8-46af-b494-523064e9b7cd" (UID: "f042ee2e-f5c8-46af-b494-523064e9b7cd"). InnerVolumeSpecName "kube-api-access-27nn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.203618 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-config" (OuterVolumeSpecName: "config") pod "f042ee2e-f5c8-46af-b494-523064e9b7cd" (UID: "f042ee2e-f5c8-46af-b494-523064e9b7cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.206752 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f042ee2e-f5c8-46af-b494-523064e9b7cd" (UID: "f042ee2e-f5c8-46af-b494-523064e9b7cd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.227956 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f042ee2e-f5c8-46af-b494-523064e9b7cd" (UID: "f042ee2e-f5c8-46af-b494-523064e9b7cd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.235967 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f042ee2e-f5c8-46af-b494-523064e9b7cd" (UID: "f042ee2e-f5c8-46af-b494-523064e9b7cd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.239185 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.239216 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.239227 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27nn4\" (UniqueName: \"kubernetes.io/projected/f042ee2e-f5c8-46af-b494-523064e9b7cd-kube-api-access-27nn4\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.239236 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.239244 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f042ee2e-f5c8-46af-b494-523064e9b7cd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.306268 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-799db77f-npk6z" event={"ID":"f042ee2e-f5c8-46af-b494-523064e9b7cd","Type":"ContainerDied","Data":"c56856aadf4f7eeecc75ff86811d67d8b26b63bdf60986a22e9c80f88bf5e164"} Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.306304 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-799db77f-npk6z" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.306344 4815 scope.go:117] "RemoveContainer" containerID="7f2eb5874d49f85f7ac50f51efe41dfd444da805af11a9738249a868e55988a6" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.318760 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf747dae-9ee6-42c5-a93d-a944f364c073","Type":"ContainerStarted","Data":"c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984"} Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.319008 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf747dae-9ee6-42c5-a93d-a944f364c073","Type":"ContainerStarted","Data":"24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8"} Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.319108 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf747dae-9ee6-42c5-a93d-a944f364c073","Type":"ContainerStarted","Data":"2dcb7ff314bbd9ce04abae6b3815a926e05ede047fa8f8121f54513b30c23c59"} Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.338406 4815 scope.go:117] "RemoveContainer" containerID="7518a815d82fdf4bf9a6d7bc1fce3eb51821b37429b22d390bc1f754166e1718" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.362881 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.362861015 podStartE2EDuration="2.362861015s" podCreationTimestamp="2025-12-05 09:30:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:30:55.348947175 +0000 UTC m=+1454.227554012" watchObservedRunningTime="2025-12-05 09:30:55.362861015 +0000 UTC m=+1454.241467852" Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.370100 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-799db77f-npk6z"] Dec 05 09:30:55 crc kubenswrapper[4815]: I1205 09:30:55.495080 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-799db77f-npk6z"] Dec 05 09:30:57 crc kubenswrapper[4815]: I1205 09:30:57.429972 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f042ee2e-f5c8-46af-b494-523064e9b7cd" path="/var/lib/kubelet/pods/f042ee2e-f5c8-46af-b494-523064e9b7cd/volumes" Dec 05 09:30:58 crc kubenswrapper[4815]: I1205 09:30:58.344741 4815 generic.go:334] "Generic (PLEG): container finished" podID="6f42cc49-79a7-494e-9f3e-da2da9da9940" containerID="41ea05b3ae7a09bbe7565edab94ed72b396859a198682ebbcb08d99b0730e0ba" exitCode=0 Dec 05 09:30:58 crc kubenswrapper[4815]: I1205 09:30:58.344812 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dxtfw" event={"ID":"6f42cc49-79a7-494e-9f3e-da2da9da9940","Type":"ContainerDied","Data":"41ea05b3ae7a09bbe7565edab94ed72b396859a198682ebbcb08d99b0730e0ba"} Dec 05 09:30:58 crc kubenswrapper[4815]: I1205 09:30:58.346118 4815 generic.go:334] "Generic (PLEG): container finished" podID="04a96446-9720-4d97-a657-42ab3826aee3" containerID="f5127d3ec099ece6f2dc2b9875ac7d19cfe4af3a2626d723607dcf76d07b7498" exitCode=0 Dec 05 09:30:58 crc kubenswrapper[4815]: I1205 09:30:58.346142 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jxrzv" event={"ID":"04a96446-9720-4d97-a657-42ab3826aee3","Type":"ContainerDied","Data":"f5127d3ec099ece6f2dc2b9875ac7d19cfe4af3a2626d723607dcf76d07b7498"} Dec 05 09:30:58 crc kubenswrapper[4815]: I1205 09:30:58.705564 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 09:30:58 crc kubenswrapper[4815]: I1205 09:30:58.705624 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 09:30:59 crc kubenswrapper[4815]: I1205 09:30:59.817523 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:30:59 crc kubenswrapper[4815]: I1205 09:30:59.822868 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.094219 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-combined-ca-bundle\") pod \"6f42cc49-79a7-494e-9f3e-da2da9da9940\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.094284 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-combined-ca-bundle\") pod \"04a96446-9720-4d97-a657-42ab3826aee3\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.094324 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95zds\" (UniqueName: \"kubernetes.io/projected/6f42cc49-79a7-494e-9f3e-da2da9da9940-kube-api-access-95zds\") pod \"6f42cc49-79a7-494e-9f3e-da2da9da9940\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.094358 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-config-data\") pod \"04a96446-9720-4d97-a657-42ab3826aee3\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.094449 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-scripts\") pod \"04a96446-9720-4d97-a657-42ab3826aee3\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.094484 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlxm8\" (UniqueName: \"kubernetes.io/projected/04a96446-9720-4d97-a657-42ab3826aee3-kube-api-access-jlxm8\") pod \"04a96446-9720-4d97-a657-42ab3826aee3\" (UID: \"04a96446-9720-4d97-a657-42ab3826aee3\") " Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.094538 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-scripts\") pod \"6f42cc49-79a7-494e-9f3e-da2da9da9940\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.094593 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-config-data\") pod \"6f42cc49-79a7-494e-9f3e-da2da9da9940\" (UID: \"6f42cc49-79a7-494e-9f3e-da2da9da9940\") " Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.101038 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-scripts" (OuterVolumeSpecName: "scripts") pod "6f42cc49-79a7-494e-9f3e-da2da9da9940" (UID: "6f42cc49-79a7-494e-9f3e-da2da9da9940"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.106664 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-scripts" (OuterVolumeSpecName: "scripts") pod "04a96446-9720-4d97-a657-42ab3826aee3" (UID: "04a96446-9720-4d97-a657-42ab3826aee3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.106870 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f42cc49-79a7-494e-9f3e-da2da9da9940-kube-api-access-95zds" (OuterVolumeSpecName: "kube-api-access-95zds") pod "6f42cc49-79a7-494e-9f3e-da2da9da9940" (UID: "6f42cc49-79a7-494e-9f3e-da2da9da9940"). InnerVolumeSpecName "kube-api-access-95zds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.106921 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04a96446-9720-4d97-a657-42ab3826aee3-kube-api-access-jlxm8" (OuterVolumeSpecName: "kube-api-access-jlxm8") pod "04a96446-9720-4d97-a657-42ab3826aee3" (UID: "04a96446-9720-4d97-a657-42ab3826aee3"). InnerVolumeSpecName "kube-api-access-jlxm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.122066 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "04a96446-9720-4d97-a657-42ab3826aee3" (UID: "04a96446-9720-4d97-a657-42ab3826aee3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.122911 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-config-data" (OuterVolumeSpecName: "config-data") pod "6f42cc49-79a7-494e-9f3e-da2da9da9940" (UID: "6f42cc49-79a7-494e-9f3e-da2da9da9940"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.123674 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-config-data" (OuterVolumeSpecName: "config-data") pod "04a96446-9720-4d97-a657-42ab3826aee3" (UID: "04a96446-9720-4d97-a657-42ab3826aee3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.155438 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6f42cc49-79a7-494e-9f3e-da2da9da9940" (UID: "6f42cc49-79a7-494e-9f3e-da2da9da9940"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.196587 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.196619 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.196637 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.196648 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95zds\" (UniqueName: \"kubernetes.io/projected/6f42cc49-79a7-494e-9f3e-da2da9da9940-kube-api-access-95zds\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.196660 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.196670 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/04a96446-9720-4d97-a657-42ab3826aee3-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.196680 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlxm8\" (UniqueName: \"kubernetes.io/projected/04a96446-9720-4d97-a657-42ab3826aee3-kube-api-access-jlxm8\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.196693 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f42cc49-79a7-494e-9f3e-da2da9da9940-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.368578 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-jxrzv" event={"ID":"04a96446-9720-4d97-a657-42ab3826aee3","Type":"ContainerDied","Data":"fd6f20585712ca74858f7ae45e524f0cfba5a1c41e5bec48c250ec1096f892bc"} Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.368618 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd6f20585712ca74858f7ae45e524f0cfba5a1c41e5bec48c250ec1096f892bc" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.368645 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-jxrzv" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.370311 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dxtfw" event={"ID":"6f42cc49-79a7-494e-9f3e-da2da9da9940","Type":"ContainerDied","Data":"532361772e59eceac99147ea35ea6fb7eefff2e6443eb87a3f23fd887a199cfb"} Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.370334 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="532361772e59eceac99147ea35ea6fb7eefff2e6443eb87a3f23fd887a199cfb" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.370397 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dxtfw" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.462711 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 09:31:00 crc kubenswrapper[4815]: E1205 09:31:00.463093 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f42cc49-79a7-494e-9f3e-da2da9da9940" containerName="nova-cell1-conductor-db-sync" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.463112 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f42cc49-79a7-494e-9f3e-da2da9da9940" containerName="nova-cell1-conductor-db-sync" Dec 05 09:31:00 crc kubenswrapper[4815]: E1205 09:31:00.463127 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f042ee2e-f5c8-46af-b494-523064e9b7cd" containerName="init" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.463133 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f042ee2e-f5c8-46af-b494-523064e9b7cd" containerName="init" Dec 05 09:31:00 crc kubenswrapper[4815]: E1205 09:31:00.463148 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f042ee2e-f5c8-46af-b494-523064e9b7cd" containerName="dnsmasq-dns" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.463154 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f042ee2e-f5c8-46af-b494-523064e9b7cd" containerName="dnsmasq-dns" Dec 05 09:31:00 crc kubenswrapper[4815]: E1205 09:31:00.463162 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04a96446-9720-4d97-a657-42ab3826aee3" containerName="nova-manage" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.463169 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="04a96446-9720-4d97-a657-42ab3826aee3" containerName="nova-manage" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.463322 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f42cc49-79a7-494e-9f3e-da2da9da9940" containerName="nova-cell1-conductor-db-sync" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.463344 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="04a96446-9720-4d97-a657-42ab3826aee3" containerName="nova-manage" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.463361 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f042ee2e-f5c8-46af-b494-523064e9b7cd" containerName="dnsmasq-dns" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.463993 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.468039 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.480813 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.556400 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.556637 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="58342a1e-cd29-49e1-b0f5-815a6508c7de" containerName="nova-scheduler-scheduler" containerID="cri-o://450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376" gracePeriod=30 Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.589041 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.589695 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cf747dae-9ee6-42c5-a93d-a944f364c073" containerName="nova-metadata-log" containerID="cri-o://24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8" gracePeriod=30 Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.589802 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cf747dae-9ee6-42c5-a93d-a944f364c073" containerName="nova-metadata-metadata" containerID="cri-o://c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984" gracePeriod=30 Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.602867 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b5fcba-8ff7-4438-8b1c-960c0996a271-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"62b5fcba-8ff7-4438-8b1c-960c0996a271\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.602917 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.602980 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b5fcba-8ff7-4438-8b1c-960c0996a271-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"62b5fcba-8ff7-4438-8b1c-960c0996a271\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.603060 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbzct\" (UniqueName: \"kubernetes.io/projected/62b5fcba-8ff7-4438-8b1c-960c0996a271-kube-api-access-sbzct\") pod \"nova-cell1-conductor-0\" (UID: \"62b5fcba-8ff7-4438-8b1c-960c0996a271\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.603143 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="90293839-ab90-42b5-ba19-880e4fa44624" containerName="nova-api-log" containerID="cri-o://cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d" gracePeriod=30 Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.603285 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="90293839-ab90-42b5-ba19-880e4fa44624" containerName="nova-api-api" containerID="cri-o://631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42" gracePeriod=30 Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.704624 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b5fcba-8ff7-4438-8b1c-960c0996a271-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"62b5fcba-8ff7-4438-8b1c-960c0996a271\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.704730 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b5fcba-8ff7-4438-8b1c-960c0996a271-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"62b5fcba-8ff7-4438-8b1c-960c0996a271\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.704815 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbzct\" (UniqueName: \"kubernetes.io/projected/62b5fcba-8ff7-4438-8b1c-960c0996a271-kube-api-access-sbzct\") pod \"nova-cell1-conductor-0\" (UID: \"62b5fcba-8ff7-4438-8b1c-960c0996a271\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.710802 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62b5fcba-8ff7-4438-8b1c-960c0996a271-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"62b5fcba-8ff7-4438-8b1c-960c0996a271\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.713017 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62b5fcba-8ff7-4438-8b1c-960c0996a271-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"62b5fcba-8ff7-4438-8b1c-960c0996a271\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.729615 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbzct\" (UniqueName: \"kubernetes.io/projected/62b5fcba-8ff7-4438-8b1c-960c0996a271-kube-api-access-sbzct\") pod \"nova-cell1-conductor-0\" (UID: \"62b5fcba-8ff7-4438-8b1c-960c0996a271\") " pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:00 crc kubenswrapper[4815]: I1205 09:31:00.781678 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.350882 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.387104 4815 generic.go:334] "Generic (PLEG): container finished" podID="cf747dae-9ee6-42c5-a93d-a944f364c073" containerID="c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984" exitCode=0 Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.387140 4815 generic.go:334] "Generic (PLEG): container finished" podID="cf747dae-9ee6-42c5-a93d-a944f364c073" containerID="24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8" exitCode=143 Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.387183 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf747dae-9ee6-42c5-a93d-a944f364c073","Type":"ContainerDied","Data":"c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984"} Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.387210 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf747dae-9ee6-42c5-a93d-a944f364c073","Type":"ContainerDied","Data":"24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8"} Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.387220 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cf747dae-9ee6-42c5-a93d-a944f364c073","Type":"ContainerDied","Data":"2dcb7ff314bbd9ce04abae6b3815a926e05ede047fa8f8121f54513b30c23c59"} Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.387234 4815 scope.go:117] "RemoveContainer" containerID="c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.387356 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.392907 4815 generic.go:334] "Generic (PLEG): container finished" podID="90293839-ab90-42b5-ba19-880e4fa44624" containerID="cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d" exitCode=143 Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.392952 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90293839-ab90-42b5-ba19-880e4fa44624","Type":"ContainerDied","Data":"cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d"} Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.430428 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf747dae-9ee6-42c5-a93d-a944f364c073-logs\") pod \"cf747dae-9ee6-42c5-a93d-a944f364c073\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.430472 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-config-data\") pod \"cf747dae-9ee6-42c5-a93d-a944f364c073\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.430534 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-nova-metadata-tls-certs\") pod \"cf747dae-9ee6-42c5-a93d-a944f364c073\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.430570 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4tmb\" (UniqueName: \"kubernetes.io/projected/cf747dae-9ee6-42c5-a93d-a944f364c073-kube-api-access-b4tmb\") pod \"cf747dae-9ee6-42c5-a93d-a944f364c073\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.430607 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-combined-ca-bundle\") pod \"cf747dae-9ee6-42c5-a93d-a944f364c073\" (UID: \"cf747dae-9ee6-42c5-a93d-a944f364c073\") " Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.431101 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf747dae-9ee6-42c5-a93d-a944f364c073-logs" (OuterVolumeSpecName: "logs") pod "cf747dae-9ee6-42c5-a93d-a944f364c073" (UID: "cf747dae-9ee6-42c5-a93d-a944f364c073"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.431648 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cf747dae-9ee6-42c5-a93d-a944f364c073-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.437028 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf747dae-9ee6-42c5-a93d-a944f364c073-kube-api-access-b4tmb" (OuterVolumeSpecName: "kube-api-access-b4tmb") pod "cf747dae-9ee6-42c5-a93d-a944f364c073" (UID: "cf747dae-9ee6-42c5-a93d-a944f364c073"). InnerVolumeSpecName "kube-api-access-b4tmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.447301 4815 scope.go:117] "RemoveContainer" containerID="24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.465066 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf747dae-9ee6-42c5-a93d-a944f364c073" (UID: "cf747dae-9ee6-42c5-a93d-a944f364c073"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.465996 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 05 09:31:01 crc kubenswrapper[4815]: W1205 09:31:01.474396 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62b5fcba_8ff7_4438_8b1c_960c0996a271.slice/crio-7ee3eb97b27683a8639d8118bf9eedf5f26eed2f62fb8cccef0cf4ae757e00bf WatchSource:0}: Error finding container 7ee3eb97b27683a8639d8118bf9eedf5f26eed2f62fb8cccef0cf4ae757e00bf: Status 404 returned error can't find the container with id 7ee3eb97b27683a8639d8118bf9eedf5f26eed2f62fb8cccef0cf4ae757e00bf Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.477832 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-config-data" (OuterVolumeSpecName: "config-data") pod "cf747dae-9ee6-42c5-a93d-a944f364c073" (UID: "cf747dae-9ee6-42c5-a93d-a944f364c073"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.485720 4815 scope.go:117] "RemoveContainer" containerID="c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984" Dec 05 09:31:01 crc kubenswrapper[4815]: E1205 09:31:01.486320 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984\": container with ID starting with c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984 not found: ID does not exist" containerID="c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.486382 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984"} err="failed to get container status \"c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984\": rpc error: code = NotFound desc = could not find container \"c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984\": container with ID starting with c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984 not found: ID does not exist" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.486426 4815 scope.go:117] "RemoveContainer" containerID="24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8" Dec 05 09:31:01 crc kubenswrapper[4815]: E1205 09:31:01.486745 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8\": container with ID starting with 24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8 not found: ID does not exist" containerID="24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.486785 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8"} err="failed to get container status \"24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8\": rpc error: code = NotFound desc = could not find container \"24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8\": container with ID starting with 24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8 not found: ID does not exist" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.486832 4815 scope.go:117] "RemoveContainer" containerID="c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.487088 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984"} err="failed to get container status \"c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984\": rpc error: code = NotFound desc = could not find container \"c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984\": container with ID starting with c7d95e478844cee62b663115b02026fa8e51a735256bab946e5ac3b4c5c9f984 not found: ID does not exist" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.487124 4815 scope.go:117] "RemoveContainer" containerID="24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.487338 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8"} err="failed to get container status \"24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8\": rpc error: code = NotFound desc = could not find container \"24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8\": container with ID starting with 24a11119c0bb0867c36deb4c6381f7a846a680149881900b31526a6c6f5333f8 not found: ID does not exist" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.500440 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "cf747dae-9ee6-42c5-a93d-a944f364c073" (UID: "cf747dae-9ee6-42c5-a93d-a944f364c073"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.532322 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.532359 4815 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.532374 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4tmb\" (UniqueName: \"kubernetes.io/projected/cf747dae-9ee6-42c5-a93d-a944f364c073-kube-api-access-b4tmb\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.532386 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf747dae-9ee6-42c5-a93d-a944f364c073-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.783343 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.790028 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.803495 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:01 crc kubenswrapper[4815]: E1205 09:31:01.803875 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf747dae-9ee6-42c5-a93d-a944f364c073" containerName="nova-metadata-log" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.803893 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf747dae-9ee6-42c5-a93d-a944f364c073" containerName="nova-metadata-log" Dec 05 09:31:01 crc kubenswrapper[4815]: E1205 09:31:01.803927 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf747dae-9ee6-42c5-a93d-a944f364c073" containerName="nova-metadata-metadata" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.803933 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf747dae-9ee6-42c5-a93d-a944f364c073" containerName="nova-metadata-metadata" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.804096 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf747dae-9ee6-42c5-a93d-a944f364c073" containerName="nova-metadata-log" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.804118 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf747dae-9ee6-42c5-a93d-a944f364c073" containerName="nova-metadata-metadata" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.805056 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.806713 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.808058 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.818170 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.837461 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.837531 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6qxs\" (UniqueName: \"kubernetes.io/projected/cbce32f5-8fd5-4603-9a9f-c25d389d5140-kube-api-access-w6qxs\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.837566 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.837650 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-config-data\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.837691 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbce32f5-8fd5-4603-9a9f-c25d389d5140-logs\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.938477 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-config-data\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.938553 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbce32f5-8fd5-4603-9a9f-c25d389d5140-logs\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.938609 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.938654 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6qxs\" (UniqueName: \"kubernetes.io/projected/cbce32f5-8fd5-4603-9a9f-c25d389d5140-kube-api-access-w6qxs\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.938681 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.939173 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbce32f5-8fd5-4603-9a9f-c25d389d5140-logs\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.943321 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.943684 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.944158 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-config-data\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:01 crc kubenswrapper[4815]: I1205 09:31:01.964305 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6qxs\" (UniqueName: \"kubernetes.io/projected/cbce32f5-8fd5-4603-9a9f-c25d389d5140-kube-api-access-w6qxs\") pod \"nova-metadata-0\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " pod="openstack/nova-metadata-0" Dec 05 09:31:02 crc kubenswrapper[4815]: I1205 09:31:02.124138 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:31:02 crc kubenswrapper[4815]: I1205 09:31:02.417924 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"62b5fcba-8ff7-4438-8b1c-960c0996a271","Type":"ContainerStarted","Data":"ba5b23636471eeebfac5a3e3df47b496501bb19ae73a841d8a181d47c001b1b9"} Dec 05 09:31:02 crc kubenswrapper[4815]: I1205 09:31:02.418197 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"62b5fcba-8ff7-4438-8b1c-960c0996a271","Type":"ContainerStarted","Data":"7ee3eb97b27683a8639d8118bf9eedf5f26eed2f62fb8cccef0cf4ae757e00bf"} Dec 05 09:31:02 crc kubenswrapper[4815]: I1205 09:31:02.418225 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:02 crc kubenswrapper[4815]: I1205 09:31:02.440918 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.4408930189999998 podStartE2EDuration="2.440893019s" podCreationTimestamp="2025-12-05 09:31:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:31:02.43566793 +0000 UTC m=+1461.314274797" watchObservedRunningTime="2025-12-05 09:31:02.440893019 +0000 UTC m=+1461.319499856" Dec 05 09:31:02 crc kubenswrapper[4815]: W1205 09:31:02.656192 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbce32f5_8fd5_4603_9a9f_c25d389d5140.slice/crio-73047fa93536b081f115925353ce234aa5c094955431e6e9d619ab1fabd0efc2 WatchSource:0}: Error finding container 73047fa93536b081f115925353ce234aa5c094955431e6e9d619ab1fabd0efc2: Status 404 returned error can't find the container with id 73047fa93536b081f115925353ce234aa5c094955431e6e9d619ab1fabd0efc2 Dec 05 09:31:02 crc kubenswrapper[4815]: I1205 09:31:02.657813 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.430787 4815 generic.go:334] "Generic (PLEG): container finished" podID="58342a1e-cd29-49e1-b0f5-815a6508c7de" containerID="450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376" exitCode=0 Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.453784 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf747dae-9ee6-42c5-a93d-a944f364c073" path="/var/lib/kubelet/pods/cf747dae-9ee6-42c5-a93d-a944f364c073/volumes" Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.457010 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58342a1e-cd29-49e1-b0f5-815a6508c7de","Type":"ContainerDied","Data":"450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376"} Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.457066 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cbce32f5-8fd5-4603-9a9f-c25d389d5140","Type":"ContainerStarted","Data":"65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1"} Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.457083 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cbce32f5-8fd5-4603-9a9f-c25d389d5140","Type":"ContainerStarted","Data":"09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07"} Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.457096 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cbce32f5-8fd5-4603-9a9f-c25d389d5140","Type":"ContainerStarted","Data":"73047fa93536b081f115925353ce234aa5c094955431e6e9d619ab1fabd0efc2"} Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.473146 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.4731228290000002 podStartE2EDuration="2.473122829s" podCreationTimestamp="2025-12-05 09:31:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:31:03.4667598 +0000 UTC m=+1462.345366637" watchObservedRunningTime="2025-12-05 09:31:03.473122829 +0000 UTC m=+1462.351729666" Dec 05 09:31:03 crc kubenswrapper[4815]: E1205 09:31:03.511076 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376 is running failed: container process not found" containerID="450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 09:31:03 crc kubenswrapper[4815]: E1205 09:31:03.512530 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376 is running failed: container process not found" containerID="450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 09:31:03 crc kubenswrapper[4815]: E1205 09:31:03.513544 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376 is running failed: container process not found" containerID="450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 09:31:03 crc kubenswrapper[4815]: E1205 09:31:03.513579 4815 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376 is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="58342a1e-cd29-49e1-b0f5-815a6508c7de" containerName="nova-scheduler-scheduler" Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.688943 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.793132 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-config-data\") pod \"58342a1e-cd29-49e1-b0f5-815a6508c7de\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.793262 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlxkg\" (UniqueName: \"kubernetes.io/projected/58342a1e-cd29-49e1-b0f5-815a6508c7de-kube-api-access-dlxkg\") pod \"58342a1e-cd29-49e1-b0f5-815a6508c7de\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.793337 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-combined-ca-bundle\") pod \"58342a1e-cd29-49e1-b0f5-815a6508c7de\" (UID: \"58342a1e-cd29-49e1-b0f5-815a6508c7de\") " Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.799742 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58342a1e-cd29-49e1-b0f5-815a6508c7de-kube-api-access-dlxkg" (OuterVolumeSpecName: "kube-api-access-dlxkg") pod "58342a1e-cd29-49e1-b0f5-815a6508c7de" (UID: "58342a1e-cd29-49e1-b0f5-815a6508c7de"). InnerVolumeSpecName "kube-api-access-dlxkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.821707 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58342a1e-cd29-49e1-b0f5-815a6508c7de" (UID: "58342a1e-cd29-49e1-b0f5-815a6508c7de"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.840416 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-config-data" (OuterVolumeSpecName: "config-data") pod "58342a1e-cd29-49e1-b0f5-815a6508c7de" (UID: "58342a1e-cd29-49e1-b0f5-815a6508c7de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.895184 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.895226 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlxkg\" (UniqueName: \"kubernetes.io/projected/58342a1e-cd29-49e1-b0f5-815a6508c7de-kube-api-access-dlxkg\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:03 crc kubenswrapper[4815]: I1205 09:31:03.895241 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58342a1e-cd29-49e1-b0f5-815a6508c7de-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.212142 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.300903 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-config-data\") pod \"90293839-ab90-42b5-ba19-880e4fa44624\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.335455 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-config-data" (OuterVolumeSpecName: "config-data") pod "90293839-ab90-42b5-ba19-880e4fa44624" (UID: "90293839-ab90-42b5-ba19-880e4fa44624"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.402020 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90293839-ab90-42b5-ba19-880e4fa44624-logs\") pod \"90293839-ab90-42b5-ba19-880e4fa44624\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.402071 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djd6w\" (UniqueName: \"kubernetes.io/projected/90293839-ab90-42b5-ba19-880e4fa44624-kube-api-access-djd6w\") pod \"90293839-ab90-42b5-ba19-880e4fa44624\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.402332 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-combined-ca-bundle\") pod \"90293839-ab90-42b5-ba19-880e4fa44624\" (UID: \"90293839-ab90-42b5-ba19-880e4fa44624\") " Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.402815 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.403109 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90293839-ab90-42b5-ba19-880e4fa44624-logs" (OuterVolumeSpecName: "logs") pod "90293839-ab90-42b5-ba19-880e4fa44624" (UID: "90293839-ab90-42b5-ba19-880e4fa44624"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.405582 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90293839-ab90-42b5-ba19-880e4fa44624-kube-api-access-djd6w" (OuterVolumeSpecName: "kube-api-access-djd6w") pod "90293839-ab90-42b5-ba19-880e4fa44624" (UID: "90293839-ab90-42b5-ba19-880e4fa44624"). InnerVolumeSpecName "kube-api-access-djd6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.430705 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "90293839-ab90-42b5-ba19-880e4fa44624" (UID: "90293839-ab90-42b5-ba19-880e4fa44624"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.442104 4815 generic.go:334] "Generic (PLEG): container finished" podID="90293839-ab90-42b5-ba19-880e4fa44624" containerID="631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42" exitCode=0 Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.442155 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90293839-ab90-42b5-ba19-880e4fa44624","Type":"ContainerDied","Data":"631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42"} Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.442181 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"90293839-ab90-42b5-ba19-880e4fa44624","Type":"ContainerDied","Data":"8b160291575a4af50952ffe2dde448295cbc76ccfeb9e976b2489dc454c552ea"} Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.442197 4815 scope.go:117] "RemoveContainer" containerID="631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.442302 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.451093 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"58342a1e-cd29-49e1-b0f5-815a6508c7de","Type":"ContainerDied","Data":"d003a2068447d5e5f7181471d830e73bca83b86886416b5edac8cfe97150e4ff"} Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.451349 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.483606 4815 scope.go:117] "RemoveContainer" containerID="cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.485546 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.517484 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90293839-ab90-42b5-ba19-880e4fa44624-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.517568 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/90293839-ab90-42b5-ba19-880e4fa44624-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.517622 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djd6w\" (UniqueName: \"kubernetes.io/projected/90293839-ab90-42b5-ba19-880e4fa44624-kube-api-access-djd6w\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.521516 4815 scope.go:117] "RemoveContainer" containerID="631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.522613 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:04 crc kubenswrapper[4815]: E1205 09:31:04.523286 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42\": container with ID starting with 631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42 not found: ID does not exist" containerID="631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.523317 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42"} err="failed to get container status \"631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42\": rpc error: code = NotFound desc = could not find container \"631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42\": container with ID starting with 631b4d9b012715a13b5f762409aaee740113b02245f7fa7c39d17c21df820f42 not found: ID does not exist" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.523337 4815 scope.go:117] "RemoveContainer" containerID="cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d" Dec 05 09:31:04 crc kubenswrapper[4815]: E1205 09:31:04.523958 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d\": container with ID starting with cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d not found: ID does not exist" containerID="cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.523982 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d"} err="failed to get container status \"cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d\": rpc error: code = NotFound desc = could not find container \"cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d\": container with ID starting with cb7abf3e5a4b4ec447cb05b66b316f7e06a848f0670ed0d214d457f42c10fe0d not found: ID does not exist" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.523996 4815 scope.go:117] "RemoveContainer" containerID="450c83848777f911b5bdd8b2e05c30d0836b997bef6bd7a6a5bc34ab5bb2d376" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.535722 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.541483 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.552695 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:04 crc kubenswrapper[4815]: E1205 09:31:04.553175 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58342a1e-cd29-49e1-b0f5-815a6508c7de" containerName="nova-scheduler-scheduler" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.553197 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="58342a1e-cd29-49e1-b0f5-815a6508c7de" containerName="nova-scheduler-scheduler" Dec 05 09:31:04 crc kubenswrapper[4815]: E1205 09:31:04.553222 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90293839-ab90-42b5-ba19-880e4fa44624" containerName="nova-api-log" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.553230 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="90293839-ab90-42b5-ba19-880e4fa44624" containerName="nova-api-log" Dec 05 09:31:04 crc kubenswrapper[4815]: E1205 09:31:04.553253 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90293839-ab90-42b5-ba19-880e4fa44624" containerName="nova-api-api" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.553261 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="90293839-ab90-42b5-ba19-880e4fa44624" containerName="nova-api-api" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.553478 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="58342a1e-cd29-49e1-b0f5-815a6508c7de" containerName="nova-scheduler-scheduler" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.553515 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="90293839-ab90-42b5-ba19-880e4fa44624" containerName="nova-api-api" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.553544 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="90293839-ab90-42b5-ba19-880e4fa44624" containerName="nova-api-log" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.554623 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.560248 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.579434 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.581131 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.586196 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.605674 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.620827 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7tvc\" (UniqueName: \"kubernetes.io/projected/9f6ff30c-e92a-4387-9cbc-072009766260-kube-api-access-k7tvc\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.620923 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.620947 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-config-data\") pod \"nova-scheduler-0\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.620971 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.621091 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bjrc\" (UniqueName: \"kubernetes.io/projected/ff800a75-4eda-43b5-bdc9-93fb56bed49c-kube-api-access-5bjrc\") pod \"nova-scheduler-0\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.621138 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f6ff30c-e92a-4387-9cbc-072009766260-logs\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.621165 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-config-data\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.627141 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.722332 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.722394 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bjrc\" (UniqueName: \"kubernetes.io/projected/ff800a75-4eda-43b5-bdc9-93fb56bed49c-kube-api-access-5bjrc\") pod \"nova-scheduler-0\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.722446 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f6ff30c-e92a-4387-9cbc-072009766260-logs\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.722471 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-config-data\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.722610 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7tvc\" (UniqueName: \"kubernetes.io/projected/9f6ff30c-e92a-4387-9cbc-072009766260-kube-api-access-k7tvc\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.722679 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.722695 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-config-data\") pod \"nova-scheduler-0\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.724310 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f6ff30c-e92a-4387-9cbc-072009766260-logs\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.727304 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.727840 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.728002 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-config-data\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.728262 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-config-data\") pod \"nova-scheduler-0\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.749794 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bjrc\" (UniqueName: \"kubernetes.io/projected/ff800a75-4eda-43b5-bdc9-93fb56bed49c-kube-api-access-5bjrc\") pod \"nova-scheduler-0\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.752310 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7tvc\" (UniqueName: \"kubernetes.io/projected/9f6ff30c-e92a-4387-9cbc-072009766260-kube-api-access-k7tvc\") pod \"nova-api-0\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.886700 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:04 crc kubenswrapper[4815]: I1205 09:31:04.904074 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:31:05 crc kubenswrapper[4815]: I1205 09:31:05.408123 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:05 crc kubenswrapper[4815]: I1205 09:31:05.454259 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58342a1e-cd29-49e1-b0f5-815a6508c7de" path="/var/lib/kubelet/pods/58342a1e-cd29-49e1-b0f5-815a6508c7de/volumes" Dec 05 09:31:05 crc kubenswrapper[4815]: I1205 09:31:05.455260 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90293839-ab90-42b5-ba19-880e4fa44624" path="/var/lib/kubelet/pods/90293839-ab90-42b5-ba19-880e4fa44624/volumes" Dec 05 09:31:05 crc kubenswrapper[4815]: I1205 09:31:05.456256 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:05 crc kubenswrapper[4815]: I1205 09:31:05.468535 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff800a75-4eda-43b5-bdc9-93fb56bed49c","Type":"ContainerStarted","Data":"4b9a95390b5fa4519ece89d5a766f6fa06dd26aa3a234303743eeda16a75ae17"} Dec 05 09:31:05 crc kubenswrapper[4815]: I1205 09:31:05.470973 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f6ff30c-e92a-4387-9cbc-072009766260","Type":"ContainerStarted","Data":"5a9243504a3d196135e73c613909ed4ee0295a581bcff54874d5a8f3d4cf4500"} Dec 05 09:31:06 crc kubenswrapper[4815]: I1205 09:31:06.481954 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff800a75-4eda-43b5-bdc9-93fb56bed49c","Type":"ContainerStarted","Data":"df83ae0a7b9ba21487b3c16540d19d6f95e9a1b9a59f926fc7aa1f0ed50b566a"} Dec 05 09:31:06 crc kubenswrapper[4815]: I1205 09:31:06.485948 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f6ff30c-e92a-4387-9cbc-072009766260","Type":"ContainerStarted","Data":"dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11"} Dec 05 09:31:06 crc kubenswrapper[4815]: I1205 09:31:06.485988 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f6ff30c-e92a-4387-9cbc-072009766260","Type":"ContainerStarted","Data":"a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44"} Dec 05 09:31:06 crc kubenswrapper[4815]: I1205 09:31:06.537935 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.537907023 podStartE2EDuration="2.537907023s" podCreationTimestamp="2025-12-05 09:31:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:31:06.503202601 +0000 UTC m=+1465.381809438" watchObservedRunningTime="2025-12-05 09:31:06.537907023 +0000 UTC m=+1465.416513860" Dec 05 09:31:06 crc kubenswrapper[4815]: I1205 09:31:06.555657 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.555640625 podStartE2EDuration="2.555640625s" podCreationTimestamp="2025-12-05 09:31:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:31:06.551932635 +0000 UTC m=+1465.430539492" watchObservedRunningTime="2025-12-05 09:31:06.555640625 +0000 UTC m=+1465.434247462" Dec 05 09:31:07 crc kubenswrapper[4815]: I1205 09:31:07.124613 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 09:31:07 crc kubenswrapper[4815]: I1205 09:31:07.126049 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 09:31:09 crc kubenswrapper[4815]: I1205 09:31:09.914113 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 09:31:10 crc kubenswrapper[4815]: I1205 09:31:10.830038 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 05 09:31:12 crc kubenswrapper[4815]: I1205 09:31:12.125268 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 09:31:12 crc kubenswrapper[4815]: I1205 09:31:12.125583 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 09:31:12 crc kubenswrapper[4815]: I1205 09:31:12.569318 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 09:31:13 crc kubenswrapper[4815]: I1205 09:31:13.140718 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.176:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:31:13 crc kubenswrapper[4815]: I1205 09:31:13.140810 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.176:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 09:31:14 crc kubenswrapper[4815]: I1205 09:31:14.886909 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 09:31:14 crc kubenswrapper[4815]: I1205 09:31:14.887144 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 09:31:14 crc kubenswrapper[4815]: I1205 09:31:14.904452 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 09:31:14 crc kubenswrapper[4815]: I1205 09:31:14.934415 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 09:31:15 crc kubenswrapper[4815]: I1205 09:31:15.334853 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:31:15 crc kubenswrapper[4815]: I1205 09:31:15.335655 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="96590e06-205c-41dd-9fa3-ee99f499510c" containerName="kube-state-metrics" containerID="cri-o://a47aa43ff263c21233ba4fef44a4e7c8154161b9c3cec737a33b19d04a682aa7" gracePeriod=30 Dec 05 09:31:15 crc kubenswrapper[4815]: I1205 09:31:15.630061 4815 generic.go:334] "Generic (PLEG): container finished" podID="96590e06-205c-41dd-9fa3-ee99f499510c" containerID="a47aa43ff263c21233ba4fef44a4e7c8154161b9c3cec737a33b19d04a682aa7" exitCode=2 Dec 05 09:31:15 crc kubenswrapper[4815]: I1205 09:31:15.630558 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"96590e06-205c-41dd-9fa3-ee99f499510c","Type":"ContainerDied","Data":"a47aa43ff263c21233ba4fef44a4e7c8154161b9c3cec737a33b19d04a682aa7"} Dec 05 09:31:15 crc kubenswrapper[4815]: I1205 09:31:15.674138 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 09:31:15 crc kubenswrapper[4815]: I1205 09:31:15.803111 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 09:31:15 crc kubenswrapper[4815]: I1205 09:31:15.825822 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dmc7\" (UniqueName: \"kubernetes.io/projected/96590e06-205c-41dd-9fa3-ee99f499510c-kube-api-access-6dmc7\") pod \"96590e06-205c-41dd-9fa3-ee99f499510c\" (UID: \"96590e06-205c-41dd-9fa3-ee99f499510c\") " Dec 05 09:31:15 crc kubenswrapper[4815]: I1205 09:31:15.830957 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96590e06-205c-41dd-9fa3-ee99f499510c-kube-api-access-6dmc7" (OuterVolumeSpecName: "kube-api-access-6dmc7") pod "96590e06-205c-41dd-9fa3-ee99f499510c" (UID: "96590e06-205c-41dd-9fa3-ee99f499510c"). InnerVolumeSpecName "kube-api-access-6dmc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:15 crc kubenswrapper[4815]: I1205 09:31:15.926947 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dmc7\" (UniqueName: \"kubernetes.io/projected/96590e06-205c-41dd-9fa3-ee99f499510c-kube-api-access-6dmc7\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:15 crc kubenswrapper[4815]: I1205 09:31:15.970697 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9f6ff30c-e92a-4387-9cbc-072009766260" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:31:15 crc kubenswrapper[4815]: I1205 09:31:15.970697 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9f6ff30c-e92a-4387-9cbc-072009766260" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.177:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.536557 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.536840 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="ceilometer-central-agent" containerID="cri-o://59804dc53de4af87ec676cfdf4914a1b245b6034588ffc6f3f06542908860e2b" gracePeriod=30 Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.536970 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="proxy-httpd" containerID="cri-o://da11f3e7790c67a5bda952d7353f8a158de3061914204669a538fd291842c7d9" gracePeriod=30 Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.537023 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="sg-core" containerID="cri-o://2b5f013e7927829e18e66a813af08f73744307f0cb29bf8bf981419c6756125a" gracePeriod=30 Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.537061 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="ceilometer-notification-agent" containerID="cri-o://36b84d577da0326ee1081b671f6786c7c5e164a030fed329d1429bfd9260ff21" gracePeriod=30 Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.641414 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.641419 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"96590e06-205c-41dd-9fa3-ee99f499510c","Type":"ContainerDied","Data":"541cadfa3b0512e3021ee18d532d3baede1c0887cc7a9b131292b18b09cddc1f"} Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.641545 4815 scope.go:117] "RemoveContainer" containerID="a47aa43ff263c21233ba4fef44a4e7c8154161b9c3cec737a33b19d04a682aa7" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.764692 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.799541 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.810839 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:31:16 crc kubenswrapper[4815]: E1205 09:31:16.811261 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96590e06-205c-41dd-9fa3-ee99f499510c" containerName="kube-state-metrics" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.811279 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="96590e06-205c-41dd-9fa3-ee99f499510c" containerName="kube-state-metrics" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.811476 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="96590e06-205c-41dd-9fa3-ee99f499510c" containerName="kube-state-metrics" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.812139 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.819811 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.819839 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.824249 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.944049 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7f328c9-41c2-4d4d-9544-c22bcc026c33-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.945477 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b7f328c9-41c2-4d4d-9544-c22bcc026c33-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.946568 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7f328c9-41c2-4d4d-9544-c22bcc026c33-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:16 crc kubenswrapper[4815]: I1205 09:31:16.946755 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l47vg\" (UniqueName: \"kubernetes.io/projected/b7f328c9-41c2-4d4d-9544-c22bcc026c33-kube-api-access-l47vg\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.048570 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7f328c9-41c2-4d4d-9544-c22bcc026c33-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.048711 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b7f328c9-41c2-4d4d-9544-c22bcc026c33-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.048777 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7f328c9-41c2-4d4d-9544-c22bcc026c33-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.048816 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l47vg\" (UniqueName: \"kubernetes.io/projected/b7f328c9-41c2-4d4d-9544-c22bcc026c33-kube-api-access-l47vg\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.057502 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/b7f328c9-41c2-4d4d-9544-c22bcc026c33-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.057883 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/b7f328c9-41c2-4d4d-9544-c22bcc026c33-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.078258 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7f328c9-41c2-4d4d-9544-c22bcc026c33-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.078818 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l47vg\" (UniqueName: \"kubernetes.io/projected/b7f328c9-41c2-4d4d-9544-c22bcc026c33-kube-api-access-l47vg\") pod \"kube-state-metrics-0\" (UID: \"b7f328c9-41c2-4d4d-9544-c22bcc026c33\") " pod="openstack/kube-state-metrics-0" Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.130554 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.433514 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96590e06-205c-41dd-9fa3-ee99f499510c" path="/var/lib/kubelet/pods/96590e06-205c-41dd-9fa3-ee99f499510c/volumes" Dec 05 09:31:17 crc kubenswrapper[4815]: W1205 09:31:17.635762 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb7f328c9_41c2_4d4d_9544_c22bcc026c33.slice/crio-e9559d161e7ff90163080473d1729d9d693b6412bc1999c49de3e032756510c1 WatchSource:0}: Error finding container e9559d161e7ff90163080473d1729d9d693b6412bc1999c49de3e032756510c1: Status 404 returned error can't find the container with id e9559d161e7ff90163080473d1729d9d693b6412bc1999c49de3e032756510c1 Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.657906 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.670039 4815 generic.go:334] "Generic (PLEG): container finished" podID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerID="da11f3e7790c67a5bda952d7353f8a158de3061914204669a538fd291842c7d9" exitCode=0 Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.670278 4815 generic.go:334] "Generic (PLEG): container finished" podID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerID="2b5f013e7927829e18e66a813af08f73744307f0cb29bf8bf981419c6756125a" exitCode=2 Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.670382 4815 generic.go:334] "Generic (PLEG): container finished" podID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerID="59804dc53de4af87ec676cfdf4914a1b245b6034588ffc6f3f06542908860e2b" exitCode=0 Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.670103 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cf6b21f-8ae2-49db-a48c-1fc941793393","Type":"ContainerDied","Data":"da11f3e7790c67a5bda952d7353f8a158de3061914204669a538fd291842c7d9"} Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.670553 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cf6b21f-8ae2-49db-a48c-1fc941793393","Type":"ContainerDied","Data":"2b5f013e7927829e18e66a813af08f73744307f0cb29bf8bf981419c6756125a"} Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.670571 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cf6b21f-8ae2-49db-a48c-1fc941793393","Type":"ContainerDied","Data":"59804dc53de4af87ec676cfdf4914a1b245b6034588ffc6f3f06542908860e2b"} Dec 05 09:31:17 crc kubenswrapper[4815]: I1205 09:31:17.672813 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b7f328c9-41c2-4d4d-9544-c22bcc026c33","Type":"ContainerStarted","Data":"e9559d161e7ff90163080473d1729d9d693b6412bc1999c49de3e032756510c1"} Dec 05 09:31:18 crc kubenswrapper[4815]: I1205 09:31:18.686415 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"b7f328c9-41c2-4d4d-9544-c22bcc026c33","Type":"ContainerStarted","Data":"79a8d2feabc805cede158fe068332f4095d0135d8b05730b989a331190c75a19"} Dec 05 09:31:18 crc kubenswrapper[4815]: I1205 09:31:18.688208 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.638644 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.661434 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=5.109561451 podStartE2EDuration="5.66141212s" podCreationTimestamp="2025-12-05 09:31:16 +0000 UTC" firstStartedPulling="2025-12-05 09:31:17.639789538 +0000 UTC m=+1476.518396375" lastFinishedPulling="2025-12-05 09:31:18.191640207 +0000 UTC m=+1477.070247044" observedRunningTime="2025-12-05 09:31:18.71313106 +0000 UTC m=+1477.591737897" watchObservedRunningTime="2025-12-05 09:31:21.66141212 +0000 UTC m=+1480.540018947" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.715395 4815 generic.go:334] "Generic (PLEG): container finished" podID="3cdb4e0e-b110-4319-b2e0-c20ef14f0525" containerID="ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a" exitCode=137 Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.715458 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3cdb4e0e-b110-4319-b2e0-c20ef14f0525","Type":"ContainerDied","Data":"ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a"} Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.715523 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"3cdb4e0e-b110-4319-b2e0-c20ef14f0525","Type":"ContainerDied","Data":"610e760da897bdea1cda35a7f0587c945a55b45695dced1df510ea6f097f7e79"} Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.715542 4815 scope.go:117] "RemoveContainer" containerID="ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.715621 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.719114 4815 generic.go:334] "Generic (PLEG): container finished" podID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerID="36b84d577da0326ee1081b671f6786c7c5e164a030fed329d1429bfd9260ff21" exitCode=0 Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.719148 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cf6b21f-8ae2-49db-a48c-1fc941793393","Type":"ContainerDied","Data":"36b84d577da0326ee1081b671f6786c7c5e164a030fed329d1429bfd9260ff21"} Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.733658 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88ghq\" (UniqueName: \"kubernetes.io/projected/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-kube-api-access-88ghq\") pod \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.733846 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-config-data\") pod \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.733884 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-combined-ca-bundle\") pod \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\" (UID: \"3cdb4e0e-b110-4319-b2e0-c20ef14f0525\") " Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.761268 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-kube-api-access-88ghq" (OuterVolumeSpecName: "kube-api-access-88ghq") pod "3cdb4e0e-b110-4319-b2e0-c20ef14f0525" (UID: "3cdb4e0e-b110-4319-b2e0-c20ef14f0525"). InnerVolumeSpecName "kube-api-access-88ghq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.833674 4815 scope.go:117] "RemoveContainer" containerID="ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a" Dec 05 09:31:21 crc kubenswrapper[4815]: E1205 09:31:21.836625 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a\": container with ID starting with ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a not found: ID does not exist" containerID="ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.836669 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a"} err="failed to get container status \"ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a\": rpc error: code = NotFound desc = could not find container \"ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a\": container with ID starting with ace3ff1cd05dfed7a6c3614192efb04bf48e7e819bb77a47bdae8fd61a67ae2a not found: ID does not exist" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.837880 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88ghq\" (UniqueName: \"kubernetes.io/projected/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-kube-api-access-88ghq\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.885775 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-config-data" (OuterVolumeSpecName: "config-data") pod "3cdb4e0e-b110-4319-b2e0-c20ef14f0525" (UID: "3cdb4e0e-b110-4319-b2e0-c20ef14f0525"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.913597 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3cdb4e0e-b110-4319-b2e0-c20ef14f0525" (UID: "3cdb4e0e-b110-4319-b2e0-c20ef14f0525"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.940943 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:21 crc kubenswrapper[4815]: I1205 09:31:21.940977 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3cdb4e0e-b110-4319-b2e0-c20ef14f0525-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.002279 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.070627 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.094433 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.104472 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 09:31:22 crc kubenswrapper[4815]: E1205 09:31:22.105131 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="ceilometer-notification-agent" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.105260 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="ceilometer-notification-agent" Dec 05 09:31:22 crc kubenswrapper[4815]: E1205 09:31:22.105376 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="sg-core" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.105450 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="sg-core" Dec 05 09:31:22 crc kubenswrapper[4815]: E1205 09:31:22.105554 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cdb4e0e-b110-4319-b2e0-c20ef14f0525" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.105631 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cdb4e0e-b110-4319-b2e0-c20ef14f0525" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 09:31:22 crc kubenswrapper[4815]: E1205 09:31:22.105711 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="ceilometer-central-agent" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.105785 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="ceilometer-central-agent" Dec 05 09:31:22 crc kubenswrapper[4815]: E1205 09:31:22.105878 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="proxy-httpd" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.106078 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="proxy-httpd" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.106416 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cdb4e0e-b110-4319-b2e0-c20ef14f0525" containerName="nova-cell1-novncproxy-novncproxy" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.106536 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="sg-core" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.106621 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="ceilometer-central-agent" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.106707 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="ceilometer-notification-agent" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.106783 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" containerName="proxy-httpd" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.107756 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.111305 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.111621 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.112381 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.115650 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.133427 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.145137 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-combined-ca-bundle\") pod \"5cf6b21f-8ae2-49db-a48c-1fc941793393\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.145214 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-sg-core-conf-yaml\") pod \"5cf6b21f-8ae2-49db-a48c-1fc941793393\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.145278 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9hkb\" (UniqueName: \"kubernetes.io/projected/5cf6b21f-8ae2-49db-a48c-1fc941793393-kube-api-access-x9hkb\") pod \"5cf6b21f-8ae2-49db-a48c-1fc941793393\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.145320 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-config-data\") pod \"5cf6b21f-8ae2-49db-a48c-1fc941793393\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.145344 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-scripts\") pod \"5cf6b21f-8ae2-49db-a48c-1fc941793393\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.145391 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-run-httpd\") pod \"5cf6b21f-8ae2-49db-a48c-1fc941793393\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.145439 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-log-httpd\") pod \"5cf6b21f-8ae2-49db-a48c-1fc941793393\" (UID: \"5cf6b21f-8ae2-49db-a48c-1fc941793393\") " Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.146219 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5cf6b21f-8ae2-49db-a48c-1fc941793393" (UID: "5cf6b21f-8ae2-49db-a48c-1fc941793393"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.147794 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5cf6b21f-8ae2-49db-a48c-1fc941793393" (UID: "5cf6b21f-8ae2-49db-a48c-1fc941793393"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.149919 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cf6b21f-8ae2-49db-a48c-1fc941793393-kube-api-access-x9hkb" (OuterVolumeSpecName: "kube-api-access-x9hkb") pod "5cf6b21f-8ae2-49db-a48c-1fc941793393" (UID: "5cf6b21f-8ae2-49db-a48c-1fc941793393"). InnerVolumeSpecName "kube-api-access-x9hkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.173541 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-scripts" (OuterVolumeSpecName: "scripts") pod "5cf6b21f-8ae2-49db-a48c-1fc941793393" (UID: "5cf6b21f-8ae2-49db-a48c-1fc941793393"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.176615 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.178353 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.209136 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5cf6b21f-8ae2-49db-a48c-1fc941793393" (UID: "5cf6b21f-8ae2-49db-a48c-1fc941793393"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.247665 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.247946 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.248138 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.248259 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sr8k\" (UniqueName: \"kubernetes.io/projected/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-kube-api-access-2sr8k\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.248478 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.248660 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.248683 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9hkb\" (UniqueName: \"kubernetes.io/projected/5cf6b21f-8ae2-49db-a48c-1fc941793393-kube-api-access-x9hkb\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.248697 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.248709 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.248829 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5cf6b21f-8ae2-49db-a48c-1fc941793393-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.269089 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5cf6b21f-8ae2-49db-a48c-1fc941793393" (UID: "5cf6b21f-8ae2-49db-a48c-1fc941793393"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.289632 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-config-data" (OuterVolumeSpecName: "config-data") pod "5cf6b21f-8ae2-49db-a48c-1fc941793393" (UID: "5cf6b21f-8ae2-49db-a48c-1fc941793393"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.350593 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.350687 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.350714 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.350766 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.350802 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sr8k\" (UniqueName: \"kubernetes.io/projected/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-kube-api-access-2sr8k\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.350873 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.350886 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5cf6b21f-8ae2-49db-a48c-1fc941793393-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.356557 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.357640 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.357894 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.359931 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.367357 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sr8k\" (UniqueName: \"kubernetes.io/projected/eea3bd17-9d01-4cf4-bdb2-4f5a48a25581-kube-api-access-2sr8k\") pod \"nova-cell1-novncproxy-0\" (UID: \"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581\") " pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.431079 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.729609 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5cf6b21f-8ae2-49db-a48c-1fc941793393","Type":"ContainerDied","Data":"e6efc06cd0731cc93373e5ad7bf6870a5978561ba95adf650c0b27eb2907b16d"} Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.729961 4815 scope.go:117] "RemoveContainer" containerID="da11f3e7790c67a5bda952d7353f8a158de3061914204669a538fd291842c7d9" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.729739 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.737077 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.756221 4815 scope.go:117] "RemoveContainer" containerID="2b5f013e7927829e18e66a813af08f73744307f0cb29bf8bf981419c6756125a" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.798893 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.800706 4815 scope.go:117] "RemoveContainer" containerID="36b84d577da0326ee1081b671f6786c7c5e164a030fed329d1429bfd9260ff21" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.814597 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.827678 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.833853 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.837418 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.837597 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.837705 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.859037 4815 scope.go:117] "RemoveContainer" containerID="59804dc53de4af87ec676cfdf4914a1b245b6034588ffc6f3f06542908860e2b" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.874620 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:22 crc kubenswrapper[4815]: W1205 09:31:22.914908 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeea3bd17_9d01_4cf4_bdb2_4f5a48a25581.slice/crio-65619b1fc790f91c5d382300b8991ca33a73feee4518ce150318d7a1b4f7538a WatchSource:0}: Error finding container 65619b1fc790f91c5d382300b8991ca33a73feee4518ce150318d7a1b4f7538a: Status 404 returned error can't find the container with id 65619b1fc790f91c5d382300b8991ca33a73feee4518ce150318d7a1b4f7538a Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.929257 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.978123 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-log-httpd\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.978165 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-scripts\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.978204 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.978251 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-run-httpd\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.978280 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft8qb\" (UniqueName: \"kubernetes.io/projected/99d76079-5639-44d6-9812-71c4a21427fa-kube-api-access-ft8qb\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.978304 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-config-data\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.978327 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:22 crc kubenswrapper[4815]: I1205 09:31:22.978352 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.080752 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-run-httpd\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.080825 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft8qb\" (UniqueName: \"kubernetes.io/projected/99d76079-5639-44d6-9812-71c4a21427fa-kube-api-access-ft8qb\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.080857 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-config-data\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.080883 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.080908 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.080945 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-log-httpd\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.080967 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-scripts\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.081003 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.082059 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-log-httpd\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.082535 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-run-httpd\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.088227 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.088589 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-scripts\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.090928 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.093200 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-config-data\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.096193 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.100854 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft8qb\" (UniqueName: \"kubernetes.io/projected/99d76079-5639-44d6-9812-71c4a21427fa-kube-api-access-ft8qb\") pod \"ceilometer-0\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.176383 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.431712 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cdb4e0e-b110-4319-b2e0-c20ef14f0525" path="/var/lib/kubelet/pods/3cdb4e0e-b110-4319-b2e0-c20ef14f0525/volumes" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.432460 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cf6b21f-8ae2-49db-a48c-1fc941793393" path="/var/lib/kubelet/pods/5cf6b21f-8ae2-49db-a48c-1fc941793393/volumes" Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.670174 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:23 crc kubenswrapper[4815]: W1205 09:31:23.683522 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99d76079_5639_44d6_9812_71c4a21427fa.slice/crio-1bae788833cbb212a45c09dac72e5985598f259d288055ab1a4c2940beabb37f WatchSource:0}: Error finding container 1bae788833cbb212a45c09dac72e5985598f259d288055ab1a4c2940beabb37f: Status 404 returned error can't find the container with id 1bae788833cbb212a45c09dac72e5985598f259d288055ab1a4c2940beabb37f Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.739305 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99d76079-5639-44d6-9812-71c4a21427fa","Type":"ContainerStarted","Data":"1bae788833cbb212a45c09dac72e5985598f259d288055ab1a4c2940beabb37f"} Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.740846 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581","Type":"ContainerStarted","Data":"8a8d36eef0c8781cac3111fddb977c1063d6afc2cceebc124a89cb9b1964e868"} Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.740896 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"eea3bd17-9d01-4cf4-bdb2-4f5a48a25581","Type":"ContainerStarted","Data":"65619b1fc790f91c5d382300b8991ca33a73feee4518ce150318d7a1b4f7538a"} Dec 05 09:31:23 crc kubenswrapper[4815]: I1205 09:31:23.768932 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.7689137430000001 podStartE2EDuration="1.768913743s" podCreationTimestamp="2025-12-05 09:31:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:31:23.762888593 +0000 UTC m=+1482.641495440" watchObservedRunningTime="2025-12-05 09:31:23.768913743 +0000 UTC m=+1482.647520580" Dec 05 09:31:24 crc kubenswrapper[4815]: I1205 09:31:24.751242 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99d76079-5639-44d6-9812-71c4a21427fa","Type":"ContainerStarted","Data":"22194b56b416810c75145d4b80c889d475a710d5fe6ae7d66b594f193bcd99c5"} Dec 05 09:31:24 crc kubenswrapper[4815]: I1205 09:31:24.890610 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 09:31:24 crc kubenswrapper[4815]: I1205 09:31:24.891076 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 09:31:24 crc kubenswrapper[4815]: I1205 09:31:24.894640 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 09:31:24 crc kubenswrapper[4815]: I1205 09:31:24.895901 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 09:31:25 crc kubenswrapper[4815]: I1205 09:31:25.768687 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99d76079-5639-44d6-9812-71c4a21427fa","Type":"ContainerStarted","Data":"748bbdd314e17ccacc9fe1004997a893d8b8427daca64a5e43e3c2e0b196041e"} Dec 05 09:31:25 crc kubenswrapper[4815]: I1205 09:31:25.769149 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 09:31:25 crc kubenswrapper[4815]: I1205 09:31:25.778061 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.116566 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f64966d5-qbnzb"] Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.118341 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.171696 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f64966d5-qbnzb"] Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.236887 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-sb\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.236957 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpmw2\" (UniqueName: \"kubernetes.io/projected/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-kube-api-access-qpmw2\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.237034 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-config\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.237075 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-nb\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.237142 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-dns-svc\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.339503 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-dns-svc\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.339651 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-sb\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.339680 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpmw2\" (UniqueName: \"kubernetes.io/projected/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-kube-api-access-qpmw2\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.339755 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-config\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.339795 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-nb\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.340823 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-nb\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.341532 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-dns-svc\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.342175 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-sb\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.343085 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-config\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.395909 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpmw2\" (UniqueName: \"kubernetes.io/projected/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-kube-api-access-qpmw2\") pod \"dnsmasq-dns-f64966d5-qbnzb\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:26 crc kubenswrapper[4815]: I1205 09:31:26.458280 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:27 crc kubenswrapper[4815]: I1205 09:31:27.147971 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 05 09:31:27 crc kubenswrapper[4815]: I1205 09:31:27.251264 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f64966d5-qbnzb"] Dec 05 09:31:27 crc kubenswrapper[4815]: W1205 09:31:27.258825 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd577dab3_e67f_4bc3_b96c_55e40eb7c9a7.slice/crio-6f1e32e6749c6efb54fdf481e947c39908cc0642166744d529058860ebe98ed3 WatchSource:0}: Error finding container 6f1e32e6749c6efb54fdf481e947c39908cc0642166744d529058860ebe98ed3: Status 404 returned error can't find the container with id 6f1e32e6749c6efb54fdf481e947c39908cc0642166744d529058860ebe98ed3 Dec 05 09:31:27 crc kubenswrapper[4815]: I1205 09:31:27.432108 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:27 crc kubenswrapper[4815]: I1205 09:31:27.797574 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99d76079-5639-44d6-9812-71c4a21427fa","Type":"ContainerStarted","Data":"933e9d4a62811a5570d59944e4aadfa0840fd204320b78d7dfe4bec4479dc62d"} Dec 05 09:31:27 crc kubenswrapper[4815]: I1205 09:31:27.801850 4815 generic.go:334] "Generic (PLEG): container finished" podID="d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" containerID="00dcefd610fc0d38f025a679f158ac903e490264b824ef7b651fd4e2531f6ae6" exitCode=0 Dec 05 09:31:27 crc kubenswrapper[4815]: I1205 09:31:27.803558 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" event={"ID":"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7","Type":"ContainerDied","Data":"00dcefd610fc0d38f025a679f158ac903e490264b824ef7b651fd4e2531f6ae6"} Dec 05 09:31:27 crc kubenswrapper[4815]: I1205 09:31:27.803604 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" event={"ID":"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7","Type":"ContainerStarted","Data":"6f1e32e6749c6efb54fdf481e947c39908cc0642166744d529058860ebe98ed3"} Dec 05 09:31:28 crc kubenswrapper[4815]: I1205 09:31:28.785649 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:28 crc kubenswrapper[4815]: I1205 09:31:28.828265 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99d76079-5639-44d6-9812-71c4a21427fa","Type":"ContainerStarted","Data":"e6782f02d9c02d9b233c8d2320ead9c947db6c2b488d0941ba9bb51238ce2ac4"} Dec 05 09:31:28 crc kubenswrapper[4815]: I1205 09:31:28.828813 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:31:28 crc kubenswrapper[4815]: I1205 09:31:28.836205 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" event={"ID":"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7","Type":"ContainerStarted","Data":"2e1d416e21594c96355f5509734f19c8ba57cde8b545bba1124a2d3bf5ee8131"} Dec 05 09:31:28 crc kubenswrapper[4815]: I1205 09:31:28.836334 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:28 crc kubenswrapper[4815]: I1205 09:31:28.836497 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9f6ff30c-e92a-4387-9cbc-072009766260" containerName="nova-api-log" containerID="cri-o://a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44" gracePeriod=30 Dec 05 09:31:28 crc kubenswrapper[4815]: I1205 09:31:28.836561 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9f6ff30c-e92a-4387-9cbc-072009766260" containerName="nova-api-api" containerID="cri-o://dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11" gracePeriod=30 Dec 05 09:31:28 crc kubenswrapper[4815]: I1205 09:31:28.862971 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.330595424 podStartE2EDuration="6.862953503s" podCreationTimestamp="2025-12-05 09:31:22 +0000 UTC" firstStartedPulling="2025-12-05 09:31:23.690157841 +0000 UTC m=+1482.568764678" lastFinishedPulling="2025-12-05 09:31:28.22251591 +0000 UTC m=+1487.101122757" observedRunningTime="2025-12-05 09:31:28.85719237 +0000 UTC m=+1487.735799207" watchObservedRunningTime="2025-12-05 09:31:28.862953503 +0000 UTC m=+1487.741560340" Dec 05 09:31:28 crc kubenswrapper[4815]: I1205 09:31:28.892080 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" podStartSLOduration=2.892056196 podStartE2EDuration="2.892056196s" podCreationTimestamp="2025-12-05 09:31:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:31:28.884617428 +0000 UTC m=+1487.763224305" watchObservedRunningTime="2025-12-05 09:31:28.892056196 +0000 UTC m=+1487.770663033" Dec 05 09:31:29 crc kubenswrapper[4815]: I1205 09:31:29.551804 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:29 crc kubenswrapper[4815]: I1205 09:31:29.847159 4815 generic.go:334] "Generic (PLEG): container finished" podID="9f6ff30c-e92a-4387-9cbc-072009766260" containerID="a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44" exitCode=143 Dec 05 09:31:29 crc kubenswrapper[4815]: I1205 09:31:29.847266 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f6ff30c-e92a-4387-9cbc-072009766260","Type":"ContainerDied","Data":"a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44"} Dec 05 09:31:30 crc kubenswrapper[4815]: I1205 09:31:30.855459 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="proxy-httpd" containerID="cri-o://e6782f02d9c02d9b233c8d2320ead9c947db6c2b488d0941ba9bb51238ce2ac4" gracePeriod=30 Dec 05 09:31:30 crc kubenswrapper[4815]: I1205 09:31:30.855499 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="sg-core" containerID="cri-o://933e9d4a62811a5570d59944e4aadfa0840fd204320b78d7dfe4bec4479dc62d" gracePeriod=30 Dec 05 09:31:30 crc kubenswrapper[4815]: I1205 09:31:30.855509 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="ceilometer-notification-agent" containerID="cri-o://748bbdd314e17ccacc9fe1004997a893d8b8427daca64a5e43e3c2e0b196041e" gracePeriod=30 Dec 05 09:31:30 crc kubenswrapper[4815]: I1205 09:31:30.855414 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="ceilometer-central-agent" containerID="cri-o://22194b56b416810c75145d4b80c889d475a710d5fe6ae7d66b594f193bcd99c5" gracePeriod=30 Dec 05 09:31:31 crc kubenswrapper[4815]: I1205 09:31:31.870546 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99d76079-5639-44d6-9812-71c4a21427fa","Type":"ContainerDied","Data":"e6782f02d9c02d9b233c8d2320ead9c947db6c2b488d0941ba9bb51238ce2ac4"} Dec 05 09:31:31 crc kubenswrapper[4815]: I1205 09:31:31.870582 4815 generic.go:334] "Generic (PLEG): container finished" podID="99d76079-5639-44d6-9812-71c4a21427fa" containerID="e6782f02d9c02d9b233c8d2320ead9c947db6c2b488d0941ba9bb51238ce2ac4" exitCode=0 Dec 05 09:31:31 crc kubenswrapper[4815]: I1205 09:31:31.870876 4815 generic.go:334] "Generic (PLEG): container finished" podID="99d76079-5639-44d6-9812-71c4a21427fa" containerID="933e9d4a62811a5570d59944e4aadfa0840fd204320b78d7dfe4bec4479dc62d" exitCode=2 Dec 05 09:31:31 crc kubenswrapper[4815]: I1205 09:31:31.870896 4815 generic.go:334] "Generic (PLEG): container finished" podID="99d76079-5639-44d6-9812-71c4a21427fa" containerID="748bbdd314e17ccacc9fe1004997a893d8b8427daca64a5e43e3c2e0b196041e" exitCode=0 Dec 05 09:31:31 crc kubenswrapper[4815]: I1205 09:31:31.870918 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99d76079-5639-44d6-9812-71c4a21427fa","Type":"ContainerDied","Data":"933e9d4a62811a5570d59944e4aadfa0840fd204320b78d7dfe4bec4479dc62d"} Dec 05 09:31:31 crc kubenswrapper[4815]: I1205 09:31:31.870937 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99d76079-5639-44d6-9812-71c4a21427fa","Type":"ContainerDied","Data":"748bbdd314e17ccacc9fe1004997a893d8b8427daca64a5e43e3c2e0b196041e"} Dec 05 09:31:32 crc kubenswrapper[4815]: I1205 09:31:32.431294 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:32 crc kubenswrapper[4815]: I1205 09:31:32.461393 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:32 crc kubenswrapper[4815]: I1205 09:31:32.750644 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:32 crc kubenswrapper[4815]: I1205 09:31:32.889806 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-config-data\") pod \"9f6ff30c-e92a-4387-9cbc-072009766260\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " Dec 05 09:31:32 crc kubenswrapper[4815]: I1205 09:31:32.889987 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7tvc\" (UniqueName: \"kubernetes.io/projected/9f6ff30c-e92a-4387-9cbc-072009766260-kube-api-access-k7tvc\") pod \"9f6ff30c-e92a-4387-9cbc-072009766260\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " Dec 05 09:31:32 crc kubenswrapper[4815]: I1205 09:31:32.890037 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f6ff30c-e92a-4387-9cbc-072009766260-logs\") pod \"9f6ff30c-e92a-4387-9cbc-072009766260\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " Dec 05 09:31:32 crc kubenswrapper[4815]: I1205 09:31:32.890082 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-combined-ca-bundle\") pod \"9f6ff30c-e92a-4387-9cbc-072009766260\" (UID: \"9f6ff30c-e92a-4387-9cbc-072009766260\") " Dec 05 09:31:32 crc kubenswrapper[4815]: I1205 09:31:32.901775 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9f6ff30c-e92a-4387-9cbc-072009766260-logs" (OuterVolumeSpecName: "logs") pod "9f6ff30c-e92a-4387-9cbc-072009766260" (UID: "9f6ff30c-e92a-4387-9cbc-072009766260"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:32 crc kubenswrapper[4815]: I1205 09:31:32.908709 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f6ff30c-e92a-4387-9cbc-072009766260-kube-api-access-k7tvc" (OuterVolumeSpecName: "kube-api-access-k7tvc") pod "9f6ff30c-e92a-4387-9cbc-072009766260" (UID: "9f6ff30c-e92a-4387-9cbc-072009766260"). InnerVolumeSpecName "kube-api-access-k7tvc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:32 crc kubenswrapper[4815]: I1205 09:31:32.993808 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7tvc\" (UniqueName: \"kubernetes.io/projected/9f6ff30c-e92a-4387-9cbc-072009766260-kube-api-access-k7tvc\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:32 crc kubenswrapper[4815]: I1205 09:31:32.994019 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9f6ff30c-e92a-4387-9cbc-072009766260-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.007175 4815 generic.go:334] "Generic (PLEG): container finished" podID="99d76079-5639-44d6-9812-71c4a21427fa" containerID="22194b56b416810c75145d4b80c889d475a710d5fe6ae7d66b594f193bcd99c5" exitCode=0 Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.007263 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99d76079-5639-44d6-9812-71c4a21427fa","Type":"ContainerDied","Data":"22194b56b416810c75145d4b80c889d475a710d5fe6ae7d66b594f193bcd99c5"} Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.019398 4815 generic.go:334] "Generic (PLEG): container finished" podID="9f6ff30c-e92a-4387-9cbc-072009766260" containerID="dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11" exitCode=0 Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.019650 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.019672 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f6ff30c-e92a-4387-9cbc-072009766260","Type":"ContainerDied","Data":"dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11"} Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.020065 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9f6ff30c-e92a-4387-9cbc-072009766260","Type":"ContainerDied","Data":"5a9243504a3d196135e73c613909ed4ee0295a581bcff54874d5a8f3d4cf4500"} Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.020134 4815 scope.go:117] "RemoveContainer" containerID="dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.023479 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f6ff30c-e92a-4387-9cbc-072009766260" (UID: "9f6ff30c-e92a-4387-9cbc-072009766260"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.033734 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-config-data" (OuterVolumeSpecName: "config-data") pod "9f6ff30c-e92a-4387-9cbc-072009766260" (UID: "9f6ff30c-e92a-4387-9cbc-072009766260"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.055569 4815 scope.go:117] "RemoveContainer" containerID="a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.095854 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.095881 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f6ff30c-e92a-4387-9cbc-072009766260-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.120830 4815 scope.go:117] "RemoveContainer" containerID="dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11" Dec 05 09:31:33 crc kubenswrapper[4815]: E1205 09:31:33.124797 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11\": container with ID starting with dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11 not found: ID does not exist" containerID="dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.124843 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11"} err="failed to get container status \"dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11\": rpc error: code = NotFound desc = could not find container \"dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11\": container with ID starting with dd96f4dbaa652984a2790c86e89ef4498ab40197920ab33947024077fdd79e11 not found: ID does not exist" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.124871 4815 scope.go:117] "RemoveContainer" containerID="a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44" Dec 05 09:31:33 crc kubenswrapper[4815]: E1205 09:31:33.128639 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44\": container with ID starting with a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44 not found: ID does not exist" containerID="a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.128675 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44"} err="failed to get container status \"a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44\": rpc error: code = NotFound desc = could not find container \"a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44\": container with ID starting with a92daaabf9074b889b1c816262332c10803977fdad663ce824ad632a4beaba44 not found: ID does not exist" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.148854 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.182080 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.300269 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-log-httpd\") pod \"99d76079-5639-44d6-9812-71c4a21427fa\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.300359 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-scripts\") pod \"99d76079-5639-44d6-9812-71c4a21427fa\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.300385 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-config-data\") pod \"99d76079-5639-44d6-9812-71c4a21427fa\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.300468 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft8qb\" (UniqueName: \"kubernetes.io/projected/99d76079-5639-44d6-9812-71c4a21427fa-kube-api-access-ft8qb\") pod \"99d76079-5639-44d6-9812-71c4a21427fa\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.300505 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-sg-core-conf-yaml\") pod \"99d76079-5639-44d6-9812-71c4a21427fa\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.300534 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-combined-ca-bundle\") pod \"99d76079-5639-44d6-9812-71c4a21427fa\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.300585 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-run-httpd\") pod \"99d76079-5639-44d6-9812-71c4a21427fa\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.300628 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-ceilometer-tls-certs\") pod \"99d76079-5639-44d6-9812-71c4a21427fa\" (UID: \"99d76079-5639-44d6-9812-71c4a21427fa\") " Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.302085 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "99d76079-5639-44d6-9812-71c4a21427fa" (UID: "99d76079-5639-44d6-9812-71c4a21427fa"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.302166 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "99d76079-5639-44d6-9812-71c4a21427fa" (UID: "99d76079-5639-44d6-9812-71c4a21427fa"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.305108 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-scripts" (OuterVolumeSpecName: "scripts") pod "99d76079-5639-44d6-9812-71c4a21427fa" (UID: "99d76079-5639-44d6-9812-71c4a21427fa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.307452 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99d76079-5639-44d6-9812-71c4a21427fa-kube-api-access-ft8qb" (OuterVolumeSpecName: "kube-api-access-ft8qb") pod "99d76079-5639-44d6-9812-71c4a21427fa" (UID: "99d76079-5639-44d6-9812-71c4a21427fa"). InnerVolumeSpecName "kube-api-access-ft8qb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.333415 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-2k4k5"] Dec 05 09:31:33 crc kubenswrapper[4815]: E1205 09:31:33.333821 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f6ff30c-e92a-4387-9cbc-072009766260" containerName="nova-api-log" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.333841 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f6ff30c-e92a-4387-9cbc-072009766260" containerName="nova-api-log" Dec 05 09:31:33 crc kubenswrapper[4815]: E1205 09:31:33.333855 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f6ff30c-e92a-4387-9cbc-072009766260" containerName="nova-api-api" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.333862 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f6ff30c-e92a-4387-9cbc-072009766260" containerName="nova-api-api" Dec 05 09:31:33 crc kubenswrapper[4815]: E1205 09:31:33.333877 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="ceilometer-notification-agent" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.333883 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="ceilometer-notification-agent" Dec 05 09:31:33 crc kubenswrapper[4815]: E1205 09:31:33.333899 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="proxy-httpd" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.333904 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="proxy-httpd" Dec 05 09:31:33 crc kubenswrapper[4815]: E1205 09:31:33.333928 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="ceilometer-central-agent" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.333933 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="ceilometer-central-agent" Dec 05 09:31:33 crc kubenswrapper[4815]: E1205 09:31:33.333943 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="sg-core" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.333949 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="sg-core" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.336420 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="sg-core" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.336446 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="ceilometer-central-agent" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.336466 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="proxy-httpd" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.336475 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f6ff30c-e92a-4387-9cbc-072009766260" containerName="nova-api-api" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.336507 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="99d76079-5639-44d6-9812-71c4a21427fa" containerName="ceilometer-notification-agent" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.336517 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f6ff30c-e92a-4387-9cbc-072009766260" containerName="nova-api-log" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.337299 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.339674 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.341106 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.350243 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2k4k5"] Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.392728 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "99d76079-5639-44d6-9812-71c4a21427fa" (UID: "99d76079-5639-44d6-9812-71c4a21427fa"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.406777 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft8qb\" (UniqueName: \"kubernetes.io/projected/99d76079-5639-44d6-9812-71c4a21427fa-kube-api-access-ft8qb\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.406808 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.406818 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.406827 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/99d76079-5639-44d6-9812-71c4a21427fa-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.406837 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.419102 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "99d76079-5639-44d6-9812-71c4a21427fa" (UID: "99d76079-5639-44d6-9812-71c4a21427fa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.421834 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "99d76079-5639-44d6-9812-71c4a21427fa" (UID: "99d76079-5639-44d6-9812-71c4a21427fa"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.466890 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-config-data" (OuterVolumeSpecName: "config-data") pod "99d76079-5639-44d6-9812-71c4a21427fa" (UID: "99d76079-5639-44d6-9812-71c4a21427fa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.510354 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-config-data\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.510945 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zvrl\" (UniqueName: \"kubernetes.io/projected/8677ce04-f49f-4df4-ad0f-ee434e75799b-kube-api-access-2zvrl\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.511119 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-scripts\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.511622 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.512179 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.512275 4815 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.512415 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d76079-5639-44d6-9812-71c4a21427fa-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.538076 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.551220 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.566012 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.567828 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.569895 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.572934 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.573144 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.581708 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.614178 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.614284 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-config-data\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.614340 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zvrl\" (UniqueName: \"kubernetes.io/projected/8677ce04-f49f-4df4-ad0f-ee434e75799b-kube-api-access-2zvrl\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.614358 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-scripts\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.617471 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-scripts\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.617648 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-config-data\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.622349 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.632891 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zvrl\" (UniqueName: \"kubernetes.io/projected/8677ce04-f49f-4df4-ad0f-ee434e75799b-kube-api-access-2zvrl\") pod \"nova-cell1-cell-mapping-2k4k5\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.715335 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-config-data\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.715680 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.715783 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85d2d605-c909-4d87-b756-fce7a233e677-logs\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.715895 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-public-tls-certs\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.715991 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-internal-tls-certs\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.716144 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz965\" (UniqueName: \"kubernetes.io/projected/85d2d605-c909-4d87-b756-fce7a233e677-kube-api-access-lz965\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.817413 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-config-data\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.817504 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.817528 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85d2d605-c909-4d87-b756-fce7a233e677-logs\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.817549 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-public-tls-certs\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.817569 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-internal-tls-certs\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.817609 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz965\" (UniqueName: \"kubernetes.io/projected/85d2d605-c909-4d87-b756-fce7a233e677-kube-api-access-lz965\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.817775 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.818226 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85d2d605-c909-4d87-b756-fce7a233e677-logs\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.822102 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-internal-tls-certs\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.822140 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-config-data\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.822237 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.829887 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-public-tls-certs\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.841701 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz965\" (UniqueName: \"kubernetes.io/projected/85d2d605-c909-4d87-b756-fce7a233e677-kube-api-access-lz965\") pod \"nova-api-0\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " pod="openstack/nova-api-0" Dec 05 09:31:33 crc kubenswrapper[4815]: I1205 09:31:33.885957 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.041517 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.043625 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"99d76079-5639-44d6-9812-71c4a21427fa","Type":"ContainerDied","Data":"1bae788833cbb212a45c09dac72e5985598f259d288055ab1a4c2940beabb37f"} Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.043661 4815 scope.go:117] "RemoveContainer" containerID="e6782f02d9c02d9b233c8d2320ead9c947db6c2b488d0941ba9bb51238ce2ac4" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.086543 4815 scope.go:117] "RemoveContainer" containerID="933e9d4a62811a5570d59944e4aadfa0840fd204320b78d7dfe4bec4479dc62d" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.089108 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.104890 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.114569 4815 scope.go:117] "RemoveContainer" containerID="748bbdd314e17ccacc9fe1004997a893d8b8427daca64a5e43e3c2e0b196041e" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.116425 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.121387 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.124839 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.124994 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.125096 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.127044 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.156598 4815 scope.go:117] "RemoveContainer" containerID="22194b56b416810c75145d4b80c889d475a710d5fe6ae7d66b594f193bcd99c5" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.223718 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-config-data\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.223762 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9rp7\" (UniqueName: \"kubernetes.io/projected/561c4cfb-f0f8-4f00-98f7-f4690d343047-kube-api-access-s9rp7\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.223807 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-log-httpd\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.223827 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.224054 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-run-httpd\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.224222 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.224311 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-scripts\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.224330 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.310091 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2k4k5"] Dec 05 09:31:34 crc kubenswrapper[4815]: W1205 09:31:34.313023 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8677ce04_f49f_4df4_ad0f_ee434e75799b.slice/crio-f1a6192de8a4cec083060cbe8deb22f4c1fbb0ca8a0ac088148e1e0149d610b9 WatchSource:0}: Error finding container f1a6192de8a4cec083060cbe8deb22f4c1fbb0ca8a0ac088148e1e0149d610b9: Status 404 returned error can't find the container with id f1a6192de8a4cec083060cbe8deb22f4c1fbb0ca8a0ac088148e1e0149d610b9 Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.325508 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.325719 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.325815 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-scripts\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.325902 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-config-data\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.325982 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9rp7\" (UniqueName: \"kubernetes.io/projected/561c4cfb-f0f8-4f00-98f7-f4690d343047-kube-api-access-s9rp7\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.326069 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-log-httpd\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.326156 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.326271 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-run-httpd\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.326845 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-run-httpd\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.329152 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.329643 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-scripts\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.329844 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.330145 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.330410 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-log-httpd\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.332220 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-config-data\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.344173 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9rp7\" (UniqueName: \"kubernetes.io/projected/561c4cfb-f0f8-4f00-98f7-f4690d343047-kube-api-access-s9rp7\") pod \"ceilometer-0\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.448056 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.480055 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:34 crc kubenswrapper[4815]: W1205 09:31:34.494826 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85d2d605_c909_4d87_b756_fce7a233e677.slice/crio-28f2ceada0159c3461b5f2d5dd6a898aaeb05b3fe65226a29b5c55238c4f118d WatchSource:0}: Error finding container 28f2ceada0159c3461b5f2d5dd6a898aaeb05b3fe65226a29b5c55238c4f118d: Status 404 returned error can't find the container with id 28f2ceada0159c3461b5f2d5dd6a898aaeb05b3fe65226a29b5c55238c4f118d Dec 05 09:31:34 crc kubenswrapper[4815]: I1205 09:31:34.968396 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 09:31:34 crc kubenswrapper[4815]: W1205 09:31:34.970724 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod561c4cfb_f0f8_4f00_98f7_f4690d343047.slice/crio-bd99c9a2c987e5bdb7ee66c534b8c9743dc4a4cb486d08b575ab308407005379 WatchSource:0}: Error finding container bd99c9a2c987e5bdb7ee66c534b8c9743dc4a4cb486d08b575ab308407005379: Status 404 returned error can't find the container with id bd99c9a2c987e5bdb7ee66c534b8c9743dc4a4cb486d08b575ab308407005379 Dec 05 09:31:35 crc kubenswrapper[4815]: I1205 09:31:35.051323 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85d2d605-c909-4d87-b756-fce7a233e677","Type":"ContainerStarted","Data":"00f59d79f53768ab866bbd7fc7f3d51a947442e5063b5900d31021c4f1baedc8"} Dec 05 09:31:35 crc kubenswrapper[4815]: I1205 09:31:35.051399 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85d2d605-c909-4d87-b756-fce7a233e677","Type":"ContainerStarted","Data":"3cef96d92575e8674e0505193a7587c99e8c0447f2e5da910606ca6413cc741f"} Dec 05 09:31:35 crc kubenswrapper[4815]: I1205 09:31:35.051411 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85d2d605-c909-4d87-b756-fce7a233e677","Type":"ContainerStarted","Data":"28f2ceada0159c3461b5f2d5dd6a898aaeb05b3fe65226a29b5c55238c4f118d"} Dec 05 09:31:35 crc kubenswrapper[4815]: I1205 09:31:35.054000 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561c4cfb-f0f8-4f00-98f7-f4690d343047","Type":"ContainerStarted","Data":"bd99c9a2c987e5bdb7ee66c534b8c9743dc4a4cb486d08b575ab308407005379"} Dec 05 09:31:35 crc kubenswrapper[4815]: I1205 09:31:35.059069 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2k4k5" event={"ID":"8677ce04-f49f-4df4-ad0f-ee434e75799b","Type":"ContainerStarted","Data":"469428e29fd165fd4b6fa5f86a059013f338dabaa038f8df298eca9291560173"} Dec 05 09:31:35 crc kubenswrapper[4815]: I1205 09:31:35.059097 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2k4k5" event={"ID":"8677ce04-f49f-4df4-ad0f-ee434e75799b","Type":"ContainerStarted","Data":"f1a6192de8a4cec083060cbe8deb22f4c1fbb0ca8a0ac088148e1e0149d610b9"} Dec 05 09:31:35 crc kubenswrapper[4815]: I1205 09:31:35.080622 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.0805986 podStartE2EDuration="2.0805986s" podCreationTimestamp="2025-12-05 09:31:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:31:35.07306323 +0000 UTC m=+1493.951670077" watchObservedRunningTime="2025-12-05 09:31:35.0805986 +0000 UTC m=+1493.959205437" Dec 05 09:31:35 crc kubenswrapper[4815]: I1205 09:31:35.098642 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-2k4k5" podStartSLOduration=2.098616999 podStartE2EDuration="2.098616999s" podCreationTimestamp="2025-12-05 09:31:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:31:35.087680409 +0000 UTC m=+1493.966287256" watchObservedRunningTime="2025-12-05 09:31:35.098616999 +0000 UTC m=+1493.977223846" Dec 05 09:31:35 crc kubenswrapper[4815]: I1205 09:31:35.431120 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99d76079-5639-44d6-9812-71c4a21427fa" path="/var/lib/kubelet/pods/99d76079-5639-44d6-9812-71c4a21427fa/volumes" Dec 05 09:31:35 crc kubenswrapper[4815]: I1205 09:31:35.432030 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f6ff30c-e92a-4387-9cbc-072009766260" path="/var/lib/kubelet/pods/9f6ff30c-e92a-4387-9cbc-072009766260/volumes" Dec 05 09:31:36 crc kubenswrapper[4815]: I1205 09:31:36.070240 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561c4cfb-f0f8-4f00-98f7-f4690d343047","Type":"ContainerStarted","Data":"a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d"} Dec 05 09:31:36 crc kubenswrapper[4815]: I1205 09:31:36.460645 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:31:36 crc kubenswrapper[4815]: I1205 09:31:36.531338 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-f98bz"] Dec 05 09:31:36 crc kubenswrapper[4815]: I1205 09:31:36.531872 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" podUID="7584b3a2-aaea-4f2e-b694-269eaa08bbc2" containerName="dnsmasq-dns" containerID="cri-o://2129ec2bb316370947a664c7161bd798f9a3d150d45ccdad272ff5342c505d09" gracePeriod=10 Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.100553 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561c4cfb-f0f8-4f00-98f7-f4690d343047","Type":"ContainerStarted","Data":"69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4"} Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.100903 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561c4cfb-f0f8-4f00-98f7-f4690d343047","Type":"ContainerStarted","Data":"a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0"} Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.107915 4815 generic.go:334] "Generic (PLEG): container finished" podID="7584b3a2-aaea-4f2e-b694-269eaa08bbc2" containerID="2129ec2bb316370947a664c7161bd798f9a3d150d45ccdad272ff5342c505d09" exitCode=0 Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.107965 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" event={"ID":"7584b3a2-aaea-4f2e-b694-269eaa08bbc2","Type":"ContainerDied","Data":"2129ec2bb316370947a664c7161bd798f9a3d150d45ccdad272ff5342c505d09"} Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.197339 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.286329 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-dns-svc\") pod \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.286644 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-sb\") pod \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.286777 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-config\") pod \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.287017 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-nb\") pod \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.287351 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8k4h\" (UniqueName: \"kubernetes.io/projected/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-kube-api-access-l8k4h\") pod \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\" (UID: \"7584b3a2-aaea-4f2e-b694-269eaa08bbc2\") " Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.301950 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-kube-api-access-l8k4h" (OuterVolumeSpecName: "kube-api-access-l8k4h") pod "7584b3a2-aaea-4f2e-b694-269eaa08bbc2" (UID: "7584b3a2-aaea-4f2e-b694-269eaa08bbc2"). InnerVolumeSpecName "kube-api-access-l8k4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.340176 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7584b3a2-aaea-4f2e-b694-269eaa08bbc2" (UID: "7584b3a2-aaea-4f2e-b694-269eaa08bbc2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.373732 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7584b3a2-aaea-4f2e-b694-269eaa08bbc2" (UID: "7584b3a2-aaea-4f2e-b694-269eaa08bbc2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.380974 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-config" (OuterVolumeSpecName: "config") pod "7584b3a2-aaea-4f2e-b694-269eaa08bbc2" (UID: "7584b3a2-aaea-4f2e-b694-269eaa08bbc2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.381028 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7584b3a2-aaea-4f2e-b694-269eaa08bbc2" (UID: "7584b3a2-aaea-4f2e-b694-269eaa08bbc2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.389046 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.389076 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8k4h\" (UniqueName: \"kubernetes.io/projected/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-kube-api-access-l8k4h\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.389089 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.389101 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:37 crc kubenswrapper[4815]: I1205 09:31:37.389111 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7584b3a2-aaea-4f2e-b694-269eaa08bbc2-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:38 crc kubenswrapper[4815]: I1205 09:31:38.127428 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" event={"ID":"7584b3a2-aaea-4f2e-b694-269eaa08bbc2","Type":"ContainerDied","Data":"33cc355a535d54973ccf741a2c6eaeddb51b5378e284fd9126f2d69204902c0b"} Dec 05 09:31:38 crc kubenswrapper[4815]: I1205 09:31:38.128730 4815 scope.go:117] "RemoveContainer" containerID="2129ec2bb316370947a664c7161bd798f9a3d150d45ccdad272ff5342c505d09" Dec 05 09:31:38 crc kubenswrapper[4815]: I1205 09:31:38.128961 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-55cbc7dc9c-f98bz" Dec 05 09:31:38 crc kubenswrapper[4815]: I1205 09:31:38.167288 4815 scope.go:117] "RemoveContainer" containerID="b2ea99432ce086b24bb1c63de777a21b8b9837b7e8d7ac6d2082b9ed512fb66a" Dec 05 09:31:38 crc kubenswrapper[4815]: I1205 09:31:38.169990 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-f98bz"] Dec 05 09:31:38 crc kubenswrapper[4815]: I1205 09:31:38.196354 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-55cbc7dc9c-f98bz"] Dec 05 09:31:39 crc kubenswrapper[4815]: I1205 09:31:39.139967 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561c4cfb-f0f8-4f00-98f7-f4690d343047","Type":"ContainerStarted","Data":"8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540"} Dec 05 09:31:39 crc kubenswrapper[4815]: I1205 09:31:39.140641 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 09:31:39 crc kubenswrapper[4815]: I1205 09:31:39.429849 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7584b3a2-aaea-4f2e-b694-269eaa08bbc2" path="/var/lib/kubelet/pods/7584b3a2-aaea-4f2e-b694-269eaa08bbc2/volumes" Dec 05 09:31:41 crc kubenswrapper[4815]: I1205 09:31:41.161777 4815 generic.go:334] "Generic (PLEG): container finished" podID="8677ce04-f49f-4df4-ad0f-ee434e75799b" containerID="469428e29fd165fd4b6fa5f86a059013f338dabaa038f8df298eca9291560173" exitCode=0 Dec 05 09:31:41 crc kubenswrapper[4815]: I1205 09:31:41.161851 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2k4k5" event={"ID":"8677ce04-f49f-4df4-ad0f-ee434e75799b","Type":"ContainerDied","Data":"469428e29fd165fd4b6fa5f86a059013f338dabaa038f8df298eca9291560173"} Dec 05 09:31:41 crc kubenswrapper[4815]: I1205 09:31:41.194084 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.900596702 podStartE2EDuration="7.19406453s" podCreationTimestamp="2025-12-05 09:31:34 +0000 UTC" firstStartedPulling="2025-12-05 09:31:34.973256609 +0000 UTC m=+1493.851863446" lastFinishedPulling="2025-12-05 09:31:38.266724447 +0000 UTC m=+1497.145331274" observedRunningTime="2025-12-05 09:31:39.180237905 +0000 UTC m=+1498.058844792" watchObservedRunningTime="2025-12-05 09:31:41.19406453 +0000 UTC m=+1500.072671367" Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.618054 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.727317 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-combined-ca-bundle\") pod \"8677ce04-f49f-4df4-ad0f-ee434e75799b\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.727480 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zvrl\" (UniqueName: \"kubernetes.io/projected/8677ce04-f49f-4df4-ad0f-ee434e75799b-kube-api-access-2zvrl\") pod \"8677ce04-f49f-4df4-ad0f-ee434e75799b\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.727588 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-scripts\") pod \"8677ce04-f49f-4df4-ad0f-ee434e75799b\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.727673 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-config-data\") pod \"8677ce04-f49f-4df4-ad0f-ee434e75799b\" (UID: \"8677ce04-f49f-4df4-ad0f-ee434e75799b\") " Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.732834 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-scripts" (OuterVolumeSpecName: "scripts") pod "8677ce04-f49f-4df4-ad0f-ee434e75799b" (UID: "8677ce04-f49f-4df4-ad0f-ee434e75799b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.739579 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8677ce04-f49f-4df4-ad0f-ee434e75799b-kube-api-access-2zvrl" (OuterVolumeSpecName: "kube-api-access-2zvrl") pod "8677ce04-f49f-4df4-ad0f-ee434e75799b" (UID: "8677ce04-f49f-4df4-ad0f-ee434e75799b"). InnerVolumeSpecName "kube-api-access-2zvrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.758609 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-config-data" (OuterVolumeSpecName: "config-data") pod "8677ce04-f49f-4df4-ad0f-ee434e75799b" (UID: "8677ce04-f49f-4df4-ad0f-ee434e75799b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.760625 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8677ce04-f49f-4df4-ad0f-ee434e75799b" (UID: "8677ce04-f49f-4df4-ad0f-ee434e75799b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.830133 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.830194 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.830218 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8677ce04-f49f-4df4-ad0f-ee434e75799b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:42 crc kubenswrapper[4815]: I1205 09:31:42.830237 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zvrl\" (UniqueName: \"kubernetes.io/projected/8677ce04-f49f-4df4-ad0f-ee434e75799b-kube-api-access-2zvrl\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:43 crc kubenswrapper[4815]: I1205 09:31:43.190894 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2k4k5" event={"ID":"8677ce04-f49f-4df4-ad0f-ee434e75799b","Type":"ContainerDied","Data":"f1a6192de8a4cec083060cbe8deb22f4c1fbb0ca8a0ac088148e1e0149d610b9"} Dec 05 09:31:43 crc kubenswrapper[4815]: I1205 09:31:43.190935 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1a6192de8a4cec083060cbe8deb22f4c1fbb0ca8a0ac088148e1e0149d610b9" Dec 05 09:31:43 crc kubenswrapper[4815]: I1205 09:31:43.190989 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2k4k5" Dec 05 09:31:43 crc kubenswrapper[4815]: I1205 09:31:43.391714 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:43 crc kubenswrapper[4815]: I1205 09:31:43.392019 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="85d2d605-c909-4d87-b756-fce7a233e677" containerName="nova-api-api" containerID="cri-o://00f59d79f53768ab866bbd7fc7f3d51a947442e5063b5900d31021c4f1baedc8" gracePeriod=30 Dec 05 09:31:43 crc kubenswrapper[4815]: I1205 09:31:43.392082 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="85d2d605-c909-4d87-b756-fce7a233e677" containerName="nova-api-log" containerID="cri-o://3cef96d92575e8674e0505193a7587c99e8c0447f2e5da910606ca6413cc741f" gracePeriod=30 Dec 05 09:31:43 crc kubenswrapper[4815]: I1205 09:31:43.410162 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:43 crc kubenswrapper[4815]: I1205 09:31:43.411171 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ff800a75-4eda-43b5-bdc9-93fb56bed49c" containerName="nova-scheduler-scheduler" containerID="cri-o://df83ae0a7b9ba21487b3c16540d19d6f95e9a1b9a59f926fc7aa1f0ed50b566a" gracePeriod=30 Dec 05 09:31:43 crc kubenswrapper[4815]: I1205 09:31:43.451436 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:43 crc kubenswrapper[4815]: I1205 09:31:43.451689 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerName="nova-metadata-log" containerID="cri-o://09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07" gracePeriod=30 Dec 05 09:31:43 crc kubenswrapper[4815]: I1205 09:31:43.451832 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerName="nova-metadata-metadata" containerID="cri-o://65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1" gracePeriod=30 Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.230416 4815 generic.go:334] "Generic (PLEG): container finished" podID="85d2d605-c909-4d87-b756-fce7a233e677" containerID="00f59d79f53768ab866bbd7fc7f3d51a947442e5063b5900d31021c4f1baedc8" exitCode=0 Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.230725 4815 generic.go:334] "Generic (PLEG): container finished" podID="85d2d605-c909-4d87-b756-fce7a233e677" containerID="3cef96d92575e8674e0505193a7587c99e8c0447f2e5da910606ca6413cc741f" exitCode=143 Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.230773 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85d2d605-c909-4d87-b756-fce7a233e677","Type":"ContainerDied","Data":"00f59d79f53768ab866bbd7fc7f3d51a947442e5063b5900d31021c4f1baedc8"} Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.230800 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85d2d605-c909-4d87-b756-fce7a233e677","Type":"ContainerDied","Data":"3cef96d92575e8674e0505193a7587c99e8c0447f2e5da910606ca6413cc741f"} Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.230809 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85d2d605-c909-4d87-b756-fce7a233e677","Type":"ContainerDied","Data":"28f2ceada0159c3461b5f2d5dd6a898aaeb05b3fe65226a29b5c55238c4f118d"} Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.230821 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28f2ceada0159c3461b5f2d5dd6a898aaeb05b3fe65226a29b5c55238c4f118d" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.233114 4815 generic.go:334] "Generic (PLEG): container finished" podID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerID="09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07" exitCode=143 Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.233147 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cbce32f5-8fd5-4603-9a9f-c25d389d5140","Type":"ContainerDied","Data":"09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07"} Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.287741 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.458079 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-combined-ca-bundle\") pod \"85d2d605-c909-4d87-b756-fce7a233e677\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.458221 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85d2d605-c909-4d87-b756-fce7a233e677-logs\") pod \"85d2d605-c909-4d87-b756-fce7a233e677\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.458312 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-public-tls-certs\") pod \"85d2d605-c909-4d87-b756-fce7a233e677\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.458413 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-internal-tls-certs\") pod \"85d2d605-c909-4d87-b756-fce7a233e677\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.458594 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-config-data\") pod \"85d2d605-c909-4d87-b756-fce7a233e677\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.458709 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz965\" (UniqueName: \"kubernetes.io/projected/85d2d605-c909-4d87-b756-fce7a233e677-kube-api-access-lz965\") pod \"85d2d605-c909-4d87-b756-fce7a233e677\" (UID: \"85d2d605-c909-4d87-b756-fce7a233e677\") " Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.462328 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85d2d605-c909-4d87-b756-fce7a233e677-logs" (OuterVolumeSpecName: "logs") pod "85d2d605-c909-4d87-b756-fce7a233e677" (UID: "85d2d605-c909-4d87-b756-fce7a233e677"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.473769 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85d2d605-c909-4d87-b756-fce7a233e677-kube-api-access-lz965" (OuterVolumeSpecName: "kube-api-access-lz965") pod "85d2d605-c909-4d87-b756-fce7a233e677" (UID: "85d2d605-c909-4d87-b756-fce7a233e677"). InnerVolumeSpecName "kube-api-access-lz965". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.499387 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85d2d605-c909-4d87-b756-fce7a233e677" (UID: "85d2d605-c909-4d87-b756-fce7a233e677"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.502037 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-config-data" (OuterVolumeSpecName: "config-data") pod "85d2d605-c909-4d87-b756-fce7a233e677" (UID: "85d2d605-c909-4d87-b756-fce7a233e677"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.541612 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "85d2d605-c909-4d87-b756-fce7a233e677" (UID: "85d2d605-c909-4d87-b756-fce7a233e677"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.562022 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz965\" (UniqueName: \"kubernetes.io/projected/85d2d605-c909-4d87-b756-fce7a233e677-kube-api-access-lz965\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.562365 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.562640 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85d2d605-c909-4d87-b756-fce7a233e677-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.562825 4815 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.562952 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.564849 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "85d2d605-c909-4d87-b756-fce7a233e677" (UID: "85d2d605-c909-4d87-b756-fce7a233e677"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:44 crc kubenswrapper[4815]: I1205 09:31:44.664680 4815 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/85d2d605-c909-4d87-b756-fce7a233e677-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:44 crc kubenswrapper[4815]: E1205 09:31:44.908381 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="df83ae0a7b9ba21487b3c16540d19d6f95e9a1b9a59f926fc7aa1f0ed50b566a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 09:31:44 crc kubenswrapper[4815]: E1205 09:31:44.910427 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="df83ae0a7b9ba21487b3c16540d19d6f95e9a1b9a59f926fc7aa1f0ed50b566a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 09:31:44 crc kubenswrapper[4815]: E1205 09:31:44.912631 4815 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="df83ae0a7b9ba21487b3c16540d19d6f95e9a1b9a59f926fc7aa1f0ed50b566a" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 05 09:31:44 crc kubenswrapper[4815]: E1205 09:31:44.912683 4815 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="ff800a75-4eda-43b5-bdc9-93fb56bed49c" containerName="nova-scheduler-scheduler" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.244306 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.280684 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.288960 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.328249 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:45 crc kubenswrapper[4815]: E1205 09:31:45.328732 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8677ce04-f49f-4df4-ad0f-ee434e75799b" containerName="nova-manage" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.328749 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8677ce04-f49f-4df4-ad0f-ee434e75799b" containerName="nova-manage" Dec 05 09:31:45 crc kubenswrapper[4815]: E1205 09:31:45.328762 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7584b3a2-aaea-4f2e-b694-269eaa08bbc2" containerName="init" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.328768 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7584b3a2-aaea-4f2e-b694-269eaa08bbc2" containerName="init" Dec 05 09:31:45 crc kubenswrapper[4815]: E1205 09:31:45.328780 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d2d605-c909-4d87-b756-fce7a233e677" containerName="nova-api-api" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.328790 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d2d605-c909-4d87-b756-fce7a233e677" containerName="nova-api-api" Dec 05 09:31:45 crc kubenswrapper[4815]: E1205 09:31:45.328800 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85d2d605-c909-4d87-b756-fce7a233e677" containerName="nova-api-log" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.328805 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="85d2d605-c909-4d87-b756-fce7a233e677" containerName="nova-api-log" Dec 05 09:31:45 crc kubenswrapper[4815]: E1205 09:31:45.328813 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7584b3a2-aaea-4f2e-b694-269eaa08bbc2" containerName="dnsmasq-dns" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.328818 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7584b3a2-aaea-4f2e-b694-269eaa08bbc2" containerName="dnsmasq-dns" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.329017 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7584b3a2-aaea-4f2e-b694-269eaa08bbc2" containerName="dnsmasq-dns" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.329036 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8677ce04-f49f-4df4-ad0f-ee434e75799b" containerName="nova-manage" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.329044 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="85d2d605-c909-4d87-b756-fce7a233e677" containerName="nova-api-api" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.329060 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="85d2d605-c909-4d87-b756-fce7a233e677" containerName="nova-api-log" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.330166 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.334769 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.335451 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.335767 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.343219 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.429113 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85d2d605-c909-4d87-b756-fce7a233e677" path="/var/lib/kubelet/pods/85d2d605-c909-4d87-b756-fce7a233e677/volumes" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.478479 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c55b4a-b8d1-4604-ad9a-de77d453795f-logs\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.478798 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.478896 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.479040 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-config-data\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.479146 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-public-tls-certs\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.479233 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjbgb\" (UniqueName: \"kubernetes.io/projected/29c55b4a-b8d1-4604-ad9a-de77d453795f-kube-api-access-pjbgb\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.581194 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjbgb\" (UniqueName: \"kubernetes.io/projected/29c55b4a-b8d1-4604-ad9a-de77d453795f-kube-api-access-pjbgb\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.581515 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c55b4a-b8d1-4604-ad9a-de77d453795f-logs\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.581680 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.581784 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.581878 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-config-data\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.582009 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-public-tls-certs\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.582024 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c55b4a-b8d1-4604-ad9a-de77d453795f-logs\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.585714 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-public-tls-certs\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.585909 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-config-data\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.586064 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-internal-tls-certs\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.593052 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c55b4a-b8d1-4604-ad9a-de77d453795f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.597876 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjbgb\" (UniqueName: \"kubernetes.io/projected/29c55b4a-b8d1-4604-ad9a-de77d453795f-kube-api-access-pjbgb\") pod \"nova-api-0\" (UID: \"29c55b4a-b8d1-4604-ad9a-de77d453795f\") " pod="openstack/nova-api-0" Dec 05 09:31:45 crc kubenswrapper[4815]: I1205 09:31:45.646250 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 05 09:31:46 crc kubenswrapper[4815]: I1205 09:31:46.160068 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 05 09:31:46 crc kubenswrapper[4815]: W1205 09:31:46.169975 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod29c55b4a_b8d1_4604_ad9a_de77d453795f.slice/crio-5cbc0f7e8f6d06359a1f8e0e1479a672a9341cc8e500471bcf7f817747a7ddd9 WatchSource:0}: Error finding container 5cbc0f7e8f6d06359a1f8e0e1479a672a9341cc8e500471bcf7f817747a7ddd9: Status 404 returned error can't find the container with id 5cbc0f7e8f6d06359a1f8e0e1479a672a9341cc8e500471bcf7f817747a7ddd9 Dec 05 09:31:46 crc kubenswrapper[4815]: I1205 09:31:46.254884 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29c55b4a-b8d1-4604-ad9a-de77d453795f","Type":"ContainerStarted","Data":"5cbc0f7e8f6d06359a1f8e0e1479a672a9341cc8e500471bcf7f817747a7ddd9"} Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.123971 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.236283 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-nova-metadata-tls-certs\") pod \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.236662 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-config-data\") pod \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.236696 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6qxs\" (UniqueName: \"kubernetes.io/projected/cbce32f5-8fd5-4603-9a9f-c25d389d5140-kube-api-access-w6qxs\") pod \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.236770 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-combined-ca-bundle\") pod \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.236871 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbce32f5-8fd5-4603-9a9f-c25d389d5140-logs\") pod \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\" (UID: \"cbce32f5-8fd5-4603-9a9f-c25d389d5140\") " Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.237411 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbce32f5-8fd5-4603-9a9f-c25d389d5140-logs" (OuterVolumeSpecName: "logs") pod "cbce32f5-8fd5-4603-9a9f-c25d389d5140" (UID: "cbce32f5-8fd5-4603-9a9f-c25d389d5140"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.241034 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbce32f5-8fd5-4603-9a9f-c25d389d5140-kube-api-access-w6qxs" (OuterVolumeSpecName: "kube-api-access-w6qxs") pod "cbce32f5-8fd5-4603-9a9f-c25d389d5140" (UID: "cbce32f5-8fd5-4603-9a9f-c25d389d5140"). InnerVolumeSpecName "kube-api-access-w6qxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.267730 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cbce32f5-8fd5-4603-9a9f-c25d389d5140" (UID: "cbce32f5-8fd5-4603-9a9f-c25d389d5140"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.281110 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29c55b4a-b8d1-4604-ad9a-de77d453795f","Type":"ContainerStarted","Data":"ba52e1c43aea40240ab3f1c0e7b859332965c75b0c948982dea3d49d1d9f53b8"} Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.284907 4815 generic.go:334] "Generic (PLEG): container finished" podID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerID="65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1" exitCode=0 Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.285098 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cbce32f5-8fd5-4603-9a9f-c25d389d5140","Type":"ContainerDied","Data":"65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1"} Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.285175 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"cbce32f5-8fd5-4603-9a9f-c25d389d5140","Type":"ContainerDied","Data":"73047fa93536b081f115925353ce234aa5c094955431e6e9d619ab1fabd0efc2"} Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.285240 4815 scope.go:117] "RemoveContainer" containerID="65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.285346 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.286551 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-config-data" (OuterVolumeSpecName: "config-data") pod "cbce32f5-8fd5-4603-9a9f-c25d389d5140" (UID: "cbce32f5-8fd5-4603-9a9f-c25d389d5140"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.303725 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "cbce32f5-8fd5-4603-9a9f-c25d389d5140" (UID: "cbce32f5-8fd5-4603-9a9f-c25d389d5140"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.314570 4815 scope.go:117] "RemoveContainer" containerID="09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.331689 4815 scope.go:117] "RemoveContainer" containerID="65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1" Dec 05 09:31:47 crc kubenswrapper[4815]: E1205 09:31:47.332218 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1\": container with ID starting with 65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1 not found: ID does not exist" containerID="65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.332311 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1"} err="failed to get container status \"65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1\": rpc error: code = NotFound desc = could not find container \"65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1\": container with ID starting with 65c5bbf4cd5f02d3314d77c225a949a9662f622d24bacc520916df803cacbae1 not found: ID does not exist" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.332399 4815 scope.go:117] "RemoveContainer" containerID="09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07" Dec 05 09:31:47 crc kubenswrapper[4815]: E1205 09:31:47.332933 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07\": container with ID starting with 09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07 not found: ID does not exist" containerID="09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.333001 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07"} err="failed to get container status \"09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07\": rpc error: code = NotFound desc = could not find container \"09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07\": container with ID starting with 09e61587342768913abb028ae40eaaf3bd4aad59b65c26ae6687e4e6fc7e9e07 not found: ID does not exist" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.339360 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.339390 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbce32f5-8fd5-4603-9a9f-c25d389d5140-logs\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.339399 4815 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.339412 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbce32f5-8fd5-4603-9a9f-c25d389d5140-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.339436 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6qxs\" (UniqueName: \"kubernetes.io/projected/cbce32f5-8fd5-4603-9a9f-c25d389d5140-kube-api-access-w6qxs\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.677556 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.699350 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:47 crc kubenswrapper[4815]: E1205 09:31:47.712904 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbce32f5_8fd5_4603_9a9f_c25d389d5140.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbce32f5_8fd5_4603_9a9f_c25d389d5140.slice/crio-73047fa93536b081f115925353ce234aa5c094955431e6e9d619ab1fabd0efc2\": RecentStats: unable to find data in memory cache]" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.718155 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:47 crc kubenswrapper[4815]: E1205 09:31:47.718670 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerName="nova-metadata-log" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.718791 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerName="nova-metadata-log" Dec 05 09:31:47 crc kubenswrapper[4815]: E1205 09:31:47.718860 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerName="nova-metadata-metadata" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.718912 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerName="nova-metadata-metadata" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.719131 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerName="nova-metadata-metadata" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.719221 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" containerName="nova-metadata-log" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.720252 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.728111 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.728504 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.740102 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.850076 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-25h4c\" (UniqueName: \"kubernetes.io/projected/a7daef92-258a-4898-890a-44de7ad290b3-kube-api-access-25h4c\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.850398 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7daef92-258a-4898-890a-44de7ad290b3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.850480 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7daef92-258a-4898-890a-44de7ad290b3-config-data\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.850624 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7daef92-258a-4898-890a-44de7ad290b3-logs\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.850717 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7daef92-258a-4898-890a-44de7ad290b3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.952001 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7daef92-258a-4898-890a-44de7ad290b3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.952313 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7daef92-258a-4898-890a-44de7ad290b3-config-data\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.952418 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7daef92-258a-4898-890a-44de7ad290b3-logs\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.952591 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7daef92-258a-4898-890a-44de7ad290b3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.952684 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-25h4c\" (UniqueName: \"kubernetes.io/projected/a7daef92-258a-4898-890a-44de7ad290b3-kube-api-access-25h4c\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.953774 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a7daef92-258a-4898-890a-44de7ad290b3-logs\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.956664 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a7daef92-258a-4898-890a-44de7ad290b3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.957938 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a7daef92-258a-4898-890a-44de7ad290b3-config-data\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.970082 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a7daef92-258a-4898-890a-44de7ad290b3-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:47 crc kubenswrapper[4815]: I1205 09:31:47.971114 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-25h4c\" (UniqueName: \"kubernetes.io/projected/a7daef92-258a-4898-890a-44de7ad290b3-kube-api-access-25h4c\") pod \"nova-metadata-0\" (UID: \"a7daef92-258a-4898-890a-44de7ad290b3\") " pod="openstack/nova-metadata-0" Dec 05 09:31:48 crc kubenswrapper[4815]: I1205 09:31:48.047009 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 05 09:31:48 crc kubenswrapper[4815]: I1205 09:31:48.301353 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"29c55b4a-b8d1-4604-ad9a-de77d453795f","Type":"ContainerStarted","Data":"eb682e3944fb19363107e149e982f0359339a62c7084a50fe18ce5d27eefeaf2"} Dec 05 09:31:48 crc kubenswrapper[4815]: I1205 09:31:48.344061 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.344041846 podStartE2EDuration="3.344041846s" podCreationTimestamp="2025-12-05 09:31:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:31:48.328282177 +0000 UTC m=+1507.206889014" watchObservedRunningTime="2025-12-05 09:31:48.344041846 +0000 UTC m=+1507.222648683" Dec 05 09:31:48 crc kubenswrapper[4815]: W1205 09:31:48.521001 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda7daef92_258a_4898_890a_44de7ad290b3.slice/crio-c034df252296147583efb6b120ecc76a99d58459e9842a1dfe68ad31180e8a59 WatchSource:0}: Error finding container c034df252296147583efb6b120ecc76a99d58459e9842a1dfe68ad31180e8a59: Status 404 returned error can't find the container with id c034df252296147583efb6b120ecc76a99d58459e9842a1dfe68ad31180e8a59 Dec 05 09:31:48 crc kubenswrapper[4815]: I1205 09:31:48.530237 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.319433 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a7daef92-258a-4898-890a-44de7ad290b3","Type":"ContainerStarted","Data":"b0da50f6d26616eaac96f3fdff2540df7db5fb9dbad47c48c15c18af4823860f"} Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.320056 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a7daef92-258a-4898-890a-44de7ad290b3","Type":"ContainerStarted","Data":"6ba5e199cca23965fae2b8e82d65a71cd06dd11d581684dcda653684b54f9731"} Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.320075 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a7daef92-258a-4898-890a-44de7ad290b3","Type":"ContainerStarted","Data":"c034df252296147583efb6b120ecc76a99d58459e9842a1dfe68ad31180e8a59"} Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.320924 4815 generic.go:334] "Generic (PLEG): container finished" podID="ff800a75-4eda-43b5-bdc9-93fb56bed49c" containerID="df83ae0a7b9ba21487b3c16540d19d6f95e9a1b9a59f926fc7aa1f0ed50b566a" exitCode=0 Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.321125 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff800a75-4eda-43b5-bdc9-93fb56bed49c","Type":"ContainerDied","Data":"df83ae0a7b9ba21487b3c16540d19d6f95e9a1b9a59f926fc7aa1f0ed50b566a"} Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.341922 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.341903093 podStartE2EDuration="2.341903093s" podCreationTimestamp="2025-12-05 09:31:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:31:49.33691988 +0000 UTC m=+1508.215526717" watchObservedRunningTime="2025-12-05 09:31:49.341903093 +0000 UTC m=+1508.220509930" Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.409476 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.438026 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbce32f5-8fd5-4603-9a9f-c25d389d5140" path="/var/lib/kubelet/pods/cbce32f5-8fd5-4603-9a9f-c25d389d5140/volumes" Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.491878 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-combined-ca-bundle\") pod \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.491955 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bjrc\" (UniqueName: \"kubernetes.io/projected/ff800a75-4eda-43b5-bdc9-93fb56bed49c-kube-api-access-5bjrc\") pod \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.492052 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-config-data\") pod \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.498888 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff800a75-4eda-43b5-bdc9-93fb56bed49c-kube-api-access-5bjrc" (OuterVolumeSpecName: "kube-api-access-5bjrc") pod "ff800a75-4eda-43b5-bdc9-93fb56bed49c" (UID: "ff800a75-4eda-43b5-bdc9-93fb56bed49c"). InnerVolumeSpecName "kube-api-access-5bjrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:31:49 crc kubenswrapper[4815]: E1205 09:31:49.521503 4815 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-config-data podName:ff800a75-4eda-43b5-bdc9-93fb56bed49c nodeName:}" failed. No retries permitted until 2025-12-05 09:31:50.021452012 +0000 UTC m=+1508.900058849 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-config-data") pod "ff800a75-4eda-43b5-bdc9-93fb56bed49c" (UID: "ff800a75-4eda-43b5-bdc9-93fb56bed49c") : error deleting /var/lib/kubelet/pods/ff800a75-4eda-43b5-bdc9-93fb56bed49c/volume-subpaths: remove /var/lib/kubelet/pods/ff800a75-4eda-43b5-bdc9-93fb56bed49c/volume-subpaths: no such file or directory Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.523983 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff800a75-4eda-43b5-bdc9-93fb56bed49c" (UID: "ff800a75-4eda-43b5-bdc9-93fb56bed49c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.593975 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:49 crc kubenswrapper[4815]: I1205 09:31:49.594012 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bjrc\" (UniqueName: \"kubernetes.io/projected/ff800a75-4eda-43b5-bdc9-93fb56bed49c-kube-api-access-5bjrc\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.100740 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-config-data\") pod \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\" (UID: \"ff800a75-4eda-43b5-bdc9-93fb56bed49c\") " Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.109511 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-config-data" (OuterVolumeSpecName: "config-data") pod "ff800a75-4eda-43b5-bdc9-93fb56bed49c" (UID: "ff800a75-4eda-43b5-bdc9-93fb56bed49c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.201843 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff800a75-4eda-43b5-bdc9-93fb56bed49c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.331440 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ff800a75-4eda-43b5-bdc9-93fb56bed49c","Type":"ContainerDied","Data":"4b9a95390b5fa4519ece89d5a766f6fa06dd26aa3a234303743eeda16a75ae17"} Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.332243 4815 scope.go:117] "RemoveContainer" containerID="df83ae0a7b9ba21487b3c16540d19d6f95e9a1b9a59f926fc7aa1f0ed50b566a" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.331473 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.378045 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.388459 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.398212 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:50 crc kubenswrapper[4815]: E1205 09:31:50.398731 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff800a75-4eda-43b5-bdc9-93fb56bed49c" containerName="nova-scheduler-scheduler" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.398747 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff800a75-4eda-43b5-bdc9-93fb56bed49c" containerName="nova-scheduler-scheduler" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.398959 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff800a75-4eda-43b5-bdc9-93fb56bed49c" containerName="nova-scheduler-scheduler" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.399769 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.406109 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89bbf450-abe1-4daa-8060-40077543ac23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"89bbf450-abe1-4daa-8060-40077543ac23\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.406162 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ggqv\" (UniqueName: \"kubernetes.io/projected/89bbf450-abe1-4daa-8060-40077543ac23-kube-api-access-2ggqv\") pod \"nova-scheduler-0\" (UID: \"89bbf450-abe1-4daa-8060-40077543ac23\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.406194 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.406211 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89bbf450-abe1-4daa-8060-40077543ac23-config-data\") pod \"nova-scheduler-0\" (UID: \"89bbf450-abe1-4daa-8060-40077543ac23\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.406669 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.506993 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89bbf450-abe1-4daa-8060-40077543ac23-config-data\") pod \"nova-scheduler-0\" (UID: \"89bbf450-abe1-4daa-8060-40077543ac23\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.508577 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89bbf450-abe1-4daa-8060-40077543ac23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"89bbf450-abe1-4daa-8060-40077543ac23\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.509011 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ggqv\" (UniqueName: \"kubernetes.io/projected/89bbf450-abe1-4daa-8060-40077543ac23-kube-api-access-2ggqv\") pod \"nova-scheduler-0\" (UID: \"89bbf450-abe1-4daa-8060-40077543ac23\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.513570 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89bbf450-abe1-4daa-8060-40077543ac23-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"89bbf450-abe1-4daa-8060-40077543ac23\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.515099 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89bbf450-abe1-4daa-8060-40077543ac23-config-data\") pod \"nova-scheduler-0\" (UID: \"89bbf450-abe1-4daa-8060-40077543ac23\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.528801 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ggqv\" (UniqueName: \"kubernetes.io/projected/89bbf450-abe1-4daa-8060-40077543ac23-kube-api-access-2ggqv\") pod \"nova-scheduler-0\" (UID: \"89bbf450-abe1-4daa-8060-40077543ac23\") " pod="openstack/nova-scheduler-0" Dec 05 09:31:50 crc kubenswrapper[4815]: I1205 09:31:50.715877 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 05 09:31:51 crc kubenswrapper[4815]: I1205 09:31:51.165360 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 05 09:31:51 crc kubenswrapper[4815]: I1205 09:31:51.343635 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"89bbf450-abe1-4daa-8060-40077543ac23","Type":"ContainerStarted","Data":"73613df61f3616726181cb684f17d89ce1fe1cefc68b378724861a524f68d091"} Dec 05 09:31:51 crc kubenswrapper[4815]: I1205 09:31:51.440692 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff800a75-4eda-43b5-bdc9-93fb56bed49c" path="/var/lib/kubelet/pods/ff800a75-4eda-43b5-bdc9-93fb56bed49c/volumes" Dec 05 09:31:52 crc kubenswrapper[4815]: I1205 09:31:52.355022 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"89bbf450-abe1-4daa-8060-40077543ac23","Type":"ContainerStarted","Data":"1d4d49de3a6c8c733d26516eda46e267ccf63ee7c29984b5552f5fb86018dd55"} Dec 05 09:31:52 crc kubenswrapper[4815]: I1205 09:31:52.380615 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.380593004 podStartE2EDuration="2.380593004s" podCreationTimestamp="2025-12-05 09:31:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:31:52.370651089 +0000 UTC m=+1511.249257966" watchObservedRunningTime="2025-12-05 09:31:52.380593004 +0000 UTC m=+1511.259199851" Dec 05 09:31:53 crc kubenswrapper[4815]: I1205 09:31:53.048287 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 09:31:53 crc kubenswrapper[4815]: I1205 09:31:53.048601 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 05 09:31:55 crc kubenswrapper[4815]: I1205 09:31:55.647679 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 09:31:55 crc kubenswrapper[4815]: I1205 09:31:55.647987 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 05 09:31:55 crc kubenswrapper[4815]: I1205 09:31:55.717420 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 05 09:31:56 crc kubenswrapper[4815]: I1205 09:31:56.660615 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="29c55b4a-b8d1-4604-ad9a-de77d453795f" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.186:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 09:31:56 crc kubenswrapper[4815]: I1205 09:31:56.660907 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="29c55b4a-b8d1-4604-ad9a-de77d453795f" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.186:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 09:31:58 crc kubenswrapper[4815]: I1205 09:31:58.048304 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 09:31:58 crc kubenswrapper[4815]: I1205 09:31:58.049608 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 05 09:31:59 crc kubenswrapper[4815]: I1205 09:31:59.060790 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a7daef92-258a-4898-890a-44de7ad290b3" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 09:31:59 crc kubenswrapper[4815]: I1205 09:31:59.061140 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a7daef92-258a-4898-890a-44de7ad290b3" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.187:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 09:32:00 crc kubenswrapper[4815]: I1205 09:32:00.717580 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 05 09:32:00 crc kubenswrapper[4815]: I1205 09:32:00.773615 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 05 09:32:01 crc kubenswrapper[4815]: I1205 09:32:01.470195 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 05 09:32:04 crc kubenswrapper[4815]: I1205 09:32:04.458690 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 09:32:05 crc kubenswrapper[4815]: I1205 09:32:05.657688 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 09:32:05 crc kubenswrapper[4815]: I1205 09:32:05.658465 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 09:32:05 crc kubenswrapper[4815]: I1205 09:32:05.667089 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 05 09:32:05 crc kubenswrapper[4815]: I1205 09:32:05.667433 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 09:32:06 crc kubenswrapper[4815]: I1205 09:32:06.491647 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 05 09:32:06 crc kubenswrapper[4815]: I1205 09:32:06.500471 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 05 09:32:08 crc kubenswrapper[4815]: I1205 09:32:08.056137 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 09:32:08 crc kubenswrapper[4815]: I1205 09:32:08.065535 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 05 09:32:08 crc kubenswrapper[4815]: I1205 09:32:08.065851 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 09:32:08 crc kubenswrapper[4815]: I1205 09:32:08.514831 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 05 09:32:18 crc kubenswrapper[4815]: I1205 09:32:18.547876 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 09:32:19 crc kubenswrapper[4815]: I1205 09:32:19.765093 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 09:32:20 crc kubenswrapper[4815]: I1205 09:32:20.193067 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:32:20 crc kubenswrapper[4815]: I1205 09:32:20.193363 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.055505 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qdkcx"] Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.057534 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.069741 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdkcx"] Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.136244 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="852e4372-ed4f-4350-b013-ab36741cd980" containerName="rabbitmq" containerID="cri-o://c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4" gracePeriod=604795 Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.205119 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-utilities\") pod \"redhat-marketplace-qdkcx\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.205183 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-catalog-content\") pod \"redhat-marketplace-qdkcx\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.205215 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fh5ff\" (UniqueName: \"kubernetes.io/projected/ce5b001e-1094-426f-b412-6d6042c3e526-kube-api-access-fh5ff\") pod \"redhat-marketplace-qdkcx\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.307080 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fh5ff\" (UniqueName: \"kubernetes.io/projected/ce5b001e-1094-426f-b412-6d6042c3e526-kube-api-access-fh5ff\") pod \"redhat-marketplace-qdkcx\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.307239 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-utilities\") pod \"redhat-marketplace-qdkcx\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.307276 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-catalog-content\") pod \"redhat-marketplace-qdkcx\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.307917 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-catalog-content\") pod \"redhat-marketplace-qdkcx\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.308404 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-utilities\") pod \"redhat-marketplace-qdkcx\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.315006 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="852e4372-ed4f-4350-b013-ab36741cd980" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.98:5671: connect: connection refused" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.344813 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fh5ff\" (UniqueName: \"kubernetes.io/projected/ce5b001e-1094-426f-b412-6d6042c3e526-kube-api-access-fh5ff\") pod \"redhat-marketplace-qdkcx\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.378241 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.429309 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="e965503d-d60a-4ce1-9dee-b74fd5eb66a0" containerName="rabbitmq" containerID="cri-o://175e366dd5e953a14c204f9dec873e4e240ebdd20e12c52e0d7b74e0d64573cb" gracePeriod=604794 Dec 05 09:32:25 crc kubenswrapper[4815]: I1205 09:32:25.833234 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdkcx"] Dec 05 09:32:26 crc kubenswrapper[4815]: I1205 09:32:26.709263 4815 generic.go:334] "Generic (PLEG): container finished" podID="ce5b001e-1094-426f-b412-6d6042c3e526" containerID="3ce7dd083e49a3124ef884ae63672121e67c4117bfd647d9728521279427e824" exitCode=0 Dec 05 09:32:26 crc kubenswrapper[4815]: I1205 09:32:26.709587 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdkcx" event={"ID":"ce5b001e-1094-426f-b412-6d6042c3e526","Type":"ContainerDied","Data":"3ce7dd083e49a3124ef884ae63672121e67c4117bfd647d9728521279427e824"} Dec 05 09:32:26 crc kubenswrapper[4815]: I1205 09:32:26.709625 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdkcx" event={"ID":"ce5b001e-1094-426f-b412-6d6042c3e526","Type":"ContainerStarted","Data":"be0bf9b35f5cd0385ac758b6409e461731a2adbdeb2324a2bd30c38562c68175"} Dec 05 09:32:28 crc kubenswrapper[4815]: I1205 09:32:28.738574 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdkcx" event={"ID":"ce5b001e-1094-426f-b412-6d6042c3e526","Type":"ContainerStarted","Data":"d76a6a7ffea0c39b2d609264cd5957d06a26bc215956efa5fc0fc18436fa9ad7"} Dec 05 09:32:28 crc kubenswrapper[4815]: E1205 09:32:28.767266 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce5b001e_1094_426f_b412_6d6042c3e526.slice/crio-d76a6a7ffea0c39b2d609264cd5957d06a26bc215956efa5fc0fc18436fa9ad7.scope\": RecentStats: unable to find data in memory cache]" Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.360676 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-skz7s"] Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.362862 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.378501 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-skz7s"] Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.504749 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5l9n\" (UniqueName: \"kubernetes.io/projected/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-kube-api-access-f5l9n\") pod \"certified-operators-skz7s\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.505315 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-catalog-content\") pod \"certified-operators-skz7s\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.505469 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-utilities\") pod \"certified-operators-skz7s\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.607237 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-utilities\") pod \"certified-operators-skz7s\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.607371 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5l9n\" (UniqueName: \"kubernetes.io/projected/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-kube-api-access-f5l9n\") pod \"certified-operators-skz7s\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.607410 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-catalog-content\") pod \"certified-operators-skz7s\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.607723 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-utilities\") pod \"certified-operators-skz7s\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.608089 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-catalog-content\") pod \"certified-operators-skz7s\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.630607 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5l9n\" (UniqueName: \"kubernetes.io/projected/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-kube-api-access-f5l9n\") pod \"certified-operators-skz7s\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:29 crc kubenswrapper[4815]: I1205 09:32:29.686558 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:30 crc kubenswrapper[4815]: I1205 09:32:30.285193 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-skz7s"] Dec 05 09:32:30 crc kubenswrapper[4815]: I1205 09:32:30.768616 4815 generic.go:334] "Generic (PLEG): container finished" podID="ce5b001e-1094-426f-b412-6d6042c3e526" containerID="d76a6a7ffea0c39b2d609264cd5957d06a26bc215956efa5fc0fc18436fa9ad7" exitCode=0 Dec 05 09:32:30 crc kubenswrapper[4815]: I1205 09:32:30.768712 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdkcx" event={"ID":"ce5b001e-1094-426f-b412-6d6042c3e526","Type":"ContainerDied","Data":"d76a6a7ffea0c39b2d609264cd5957d06a26bc215956efa5fc0fc18436fa9ad7"} Dec 05 09:32:30 crc kubenswrapper[4815]: I1205 09:32:30.779007 4815 generic.go:334] "Generic (PLEG): container finished" podID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" containerID="d0ad33a617d68fdf4b0c4062d5a7b8145ad83fb3a8addbeb0bbdbc3087ec3d82" exitCode=0 Dec 05 09:32:30 crc kubenswrapper[4815]: I1205 09:32:30.779050 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skz7s" event={"ID":"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd","Type":"ContainerDied","Data":"d0ad33a617d68fdf4b0c4062d5a7b8145ad83fb3a8addbeb0bbdbc3087ec3d82"} Dec 05 09:32:30 crc kubenswrapper[4815]: I1205 09:32:30.779093 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skz7s" event={"ID":"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd","Type":"ContainerStarted","Data":"c43633b2e8a81172a41542caaa3e8a542b2de761b5b628ccfea439f2ae75f3b6"} Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.771656 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.833635 4815 generic.go:334] "Generic (PLEG): container finished" podID="e965503d-d60a-4ce1-9dee-b74fd5eb66a0" containerID="175e366dd5e953a14c204f9dec873e4e240ebdd20e12c52e0d7b74e0d64573cb" exitCode=0 Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.833797 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e965503d-d60a-4ce1-9dee-b74fd5eb66a0","Type":"ContainerDied","Data":"175e366dd5e953a14c204f9dec873e4e240ebdd20e12c52e0d7b74e0d64573cb"} Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.852947 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdkcx" event={"ID":"ce5b001e-1094-426f-b412-6d6042c3e526","Type":"ContainerStarted","Data":"c39db0357a6e90c90581575b6431bc8843876a1bb495f8bd5939eeab8e8976eb"} Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.877324 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"852e4372-ed4f-4350-b013-ab36741cd980\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.877416 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-config-data\") pod \"852e4372-ed4f-4350-b013-ab36741cd980\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.877471 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-server-conf\") pod \"852e4372-ed4f-4350-b013-ab36741cd980\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.877555 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-erlang-cookie\") pod \"852e4372-ed4f-4350-b013-ab36741cd980\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.877574 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhrl5\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-kube-api-access-xhrl5\") pod \"852e4372-ed4f-4350-b013-ab36741cd980\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.877614 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-confd\") pod \"852e4372-ed4f-4350-b013-ab36741cd980\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.877656 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-tls\") pod \"852e4372-ed4f-4350-b013-ab36741cd980\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.877687 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-plugins\") pod \"852e4372-ed4f-4350-b013-ab36741cd980\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.877718 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-plugins-conf\") pod \"852e4372-ed4f-4350-b013-ab36741cd980\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.877736 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/852e4372-ed4f-4350-b013-ab36741cd980-erlang-cookie-secret\") pod \"852e4372-ed4f-4350-b013-ab36741cd980\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.877777 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/852e4372-ed4f-4350-b013-ab36741cd980-pod-info\") pod \"852e4372-ed4f-4350-b013-ab36741cd980\" (UID: \"852e4372-ed4f-4350-b013-ab36741cd980\") " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.881887 4815 generic.go:334] "Generic (PLEG): container finished" podID="852e4372-ed4f-4350-b013-ab36741cd980" containerID="c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4" exitCode=0 Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.881930 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"852e4372-ed4f-4350-b013-ab36741cd980","Type":"ContainerDied","Data":"c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4"} Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.881955 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"852e4372-ed4f-4350-b013-ab36741cd980","Type":"ContainerDied","Data":"cb6b03b36955480bbb604abd7fa7a3fb73ca932bec9eb06df9a4eb0574fd1506"} Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.881984 4815 scope.go:117] "RemoveContainer" containerID="c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.882157 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.882540 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "852e4372-ed4f-4350-b013-ab36741cd980" (UID: "852e4372-ed4f-4350-b013-ab36741cd980"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.883185 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "852e4372-ed4f-4350-b013-ab36741cd980" (UID: "852e4372-ed4f-4350-b013-ab36741cd980"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.886131 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qdkcx" podStartSLOduration=2.396769436 podStartE2EDuration="6.886102972s" podCreationTimestamp="2025-12-05 09:32:25 +0000 UTC" firstStartedPulling="2025-12-05 09:32:26.711784589 +0000 UTC m=+1545.590391426" lastFinishedPulling="2025-12-05 09:32:31.201118115 +0000 UTC m=+1550.079724962" observedRunningTime="2025-12-05 09:32:31.877516834 +0000 UTC m=+1550.756123671" watchObservedRunningTime="2025-12-05 09:32:31.886102972 +0000 UTC m=+1550.764709809" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.890969 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "852e4372-ed4f-4350-b013-ab36741cd980" (UID: "852e4372-ed4f-4350-b013-ab36741cd980"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.953293 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "852e4372-ed4f-4350-b013-ab36741cd980" (UID: "852e4372-ed4f-4350-b013-ab36741cd980"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.961689 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/852e4372-ed4f-4350-b013-ab36741cd980-pod-info" (OuterVolumeSpecName: "pod-info") pod "852e4372-ed4f-4350-b013-ab36741cd980" (UID: "852e4372-ed4f-4350-b013-ab36741cd980"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.968952 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-kube-api-access-xhrl5" (OuterVolumeSpecName: "kube-api-access-xhrl5") pod "852e4372-ed4f-4350-b013-ab36741cd980" (UID: "852e4372-ed4f-4350-b013-ab36741cd980"). InnerVolumeSpecName "kube-api-access-xhrl5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.973232 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "852e4372-ed4f-4350-b013-ab36741cd980" (UID: "852e4372-ed4f-4350-b013-ab36741cd980"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.976430 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/852e4372-ed4f-4350-b013-ab36741cd980-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "852e4372-ed4f-4350-b013-ab36741cd980" (UID: "852e4372-ed4f-4350-b013-ab36741cd980"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.980050 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhrl5\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-kube-api-access-xhrl5\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.980078 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.980088 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.980119 4815 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.980133 4815 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/852e4372-ed4f-4350-b013-ab36741cd980-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.980147 4815 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/852e4372-ed4f-4350-b013-ab36741cd980-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.980170 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 05 09:32:31 crc kubenswrapper[4815]: I1205 09:32:31.980202 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.062057 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.088205 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.155286 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-config-data" (OuterVolumeSpecName: "config-data") pod "852e4372-ed4f-4350-b013-ab36741cd980" (UID: "852e4372-ed4f-4350-b013-ab36741cd980"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.185562 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.193068 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.201713 4815 scope.go:117] "RemoveContainer" containerID="fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.202111 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-server-conf" (OuterVolumeSpecName: "server-conf") pod "852e4372-ed4f-4350-b013-ab36741cd980" (UID: "852e4372-ed4f-4350-b013-ab36741cd980"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.254669 4815 scope.go:117] "RemoveContainer" containerID="c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4" Dec 05 09:32:32 crc kubenswrapper[4815]: E1205 09:32:32.260826 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4\": container with ID starting with c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4 not found: ID does not exist" containerID="c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.260948 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4"} err="failed to get container status \"c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4\": rpc error: code = NotFound desc = could not find container \"c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4\": container with ID starting with c923f4bad6bcfe1f6dbe17f46c2b41dd67aa29a5236578c46c0ffe6c608011d4 not found: ID does not exist" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.261011 4815 scope.go:117] "RemoveContainer" containerID="fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186" Dec 05 09:32:32 crc kubenswrapper[4815]: E1205 09:32:32.266765 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186\": container with ID starting with fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186 not found: ID does not exist" containerID="fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.266853 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186"} err="failed to get container status \"fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186\": rpc error: code = NotFound desc = could not find container \"fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186\": container with ID starting with fd6bce1a1630500711243d6a64e81516ce0c71b965836029861b27ca66fe8186 not found: ID does not exist" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.294252 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.294364 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-confd\") pod \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.294409 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-erlang-cookie\") pod \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.294448 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-plugins-conf\") pod \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.294471 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd2b7\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-kube-api-access-gd2b7\") pod \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.294513 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-plugins\") pod \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.294544 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-server-conf\") pod \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.294587 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-pod-info\") pod \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.294610 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-erlang-cookie-secret\") pod \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.294644 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-tls\") pod \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.295508 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-config-data\") pod \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\" (UID: \"e965503d-d60a-4ce1-9dee-b74fd5eb66a0\") " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.296525 4815 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/852e4372-ed4f-4350-b013-ab36741cd980-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.296937 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "e965503d-d60a-4ce1-9dee-b74fd5eb66a0" (UID: "e965503d-d60a-4ce1-9dee-b74fd5eb66a0"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.297611 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "e965503d-d60a-4ce1-9dee-b74fd5eb66a0" (UID: "e965503d-d60a-4ce1-9dee-b74fd5eb66a0"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.310416 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-pod-info" (OuterVolumeSpecName: "pod-info") pod "e965503d-d60a-4ce1-9dee-b74fd5eb66a0" (UID: "e965503d-d60a-4ce1-9dee-b74fd5eb66a0"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.310813 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "e965503d-d60a-4ce1-9dee-b74fd5eb66a0" (UID: "e965503d-d60a-4ce1-9dee-b74fd5eb66a0"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.312038 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "e965503d-d60a-4ce1-9dee-b74fd5eb66a0" (UID: "e965503d-d60a-4ce1-9dee-b74fd5eb66a0"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.314169 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "e965503d-d60a-4ce1-9dee-b74fd5eb66a0" (UID: "e965503d-d60a-4ce1-9dee-b74fd5eb66a0"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.322691 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-kube-api-access-gd2b7" (OuterVolumeSpecName: "kube-api-access-gd2b7") pod "e965503d-d60a-4ce1-9dee-b74fd5eb66a0" (UID: "e965503d-d60a-4ce1-9dee-b74fd5eb66a0"). InnerVolumeSpecName "kube-api-access-gd2b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.344860 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "e965503d-d60a-4ce1-9dee-b74fd5eb66a0" (UID: "e965503d-d60a-4ce1-9dee-b74fd5eb66a0"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.382338 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-config-data" (OuterVolumeSpecName: "config-data") pod "e965503d-d60a-4ce1-9dee-b74fd5eb66a0" (UID: "e965503d-d60a-4ce1-9dee-b74fd5eb66a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.388823 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "852e4372-ed4f-4350-b013-ab36741cd980" (UID: "852e4372-ed4f-4350-b013-ab36741cd980"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.397863 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.397903 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.397913 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.397925 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/852e4372-ed4f-4350-b013-ab36741cd980-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.397933 4815 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.397942 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd2b7\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-kube-api-access-gd2b7\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.397949 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.397957 4815 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-pod-info\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.397964 4815 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.397971 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.440934 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-server-conf" (OuterVolumeSpecName: "server-conf") pod "e965503d-d60a-4ce1-9dee-b74fd5eb66a0" (UID: "e965503d-d60a-4ce1-9dee-b74fd5eb66a0"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.450557 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.502548 4815 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-server-conf\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.502791 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.591321 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.605799 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.626659 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 09:32:32 crc kubenswrapper[4815]: E1205 09:32:32.627109 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="852e4372-ed4f-4350-b013-ab36741cd980" containerName="setup-container" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.627147 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="852e4372-ed4f-4350-b013-ab36741cd980" containerName="setup-container" Dec 05 09:32:32 crc kubenswrapper[4815]: E1205 09:32:32.627175 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e965503d-d60a-4ce1-9dee-b74fd5eb66a0" containerName="rabbitmq" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.627182 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e965503d-d60a-4ce1-9dee-b74fd5eb66a0" containerName="rabbitmq" Dec 05 09:32:32 crc kubenswrapper[4815]: E1205 09:32:32.627198 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e965503d-d60a-4ce1-9dee-b74fd5eb66a0" containerName="setup-container" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.627205 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e965503d-d60a-4ce1-9dee-b74fd5eb66a0" containerName="setup-container" Dec 05 09:32:32 crc kubenswrapper[4815]: E1205 09:32:32.627223 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="852e4372-ed4f-4350-b013-ab36741cd980" containerName="rabbitmq" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.627229 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="852e4372-ed4f-4350-b013-ab36741cd980" containerName="rabbitmq" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.627407 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="852e4372-ed4f-4350-b013-ab36741cd980" containerName="rabbitmq" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.627453 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e965503d-d60a-4ce1-9dee-b74fd5eb66a0" containerName="rabbitmq" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.628556 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.630686 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.630865 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.630899 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.632015 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.632396 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-lmbsv" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.632764 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.635128 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.636754 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.686222 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "e965503d-d60a-4ce1-9dee-b74fd5eb66a0" (UID: "e965503d-d60a-4ce1-9dee-b74fd5eb66a0"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.707950 4815 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/e965503d-d60a-4ce1-9dee-b74fd5eb66a0-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.810087 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/db8bc583-6bcf-40c7-875b-d5c9544af543-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.811126 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.811234 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sdqq\" (UniqueName: \"kubernetes.io/projected/db8bc583-6bcf-40c7-875b-d5c9544af543-kube-api-access-9sdqq\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.811364 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.811511 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/db8bc583-6bcf-40c7-875b-d5c9544af543-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.811606 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.811708 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.811784 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/db8bc583-6bcf-40c7-875b-d5c9544af543-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.811857 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.811965 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db8bc583-6bcf-40c7-875b-d5c9544af543-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.812050 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/db8bc583-6bcf-40c7-875b-d5c9544af543-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.892398 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"e965503d-d60a-4ce1-9dee-b74fd5eb66a0","Type":"ContainerDied","Data":"bd3a035bbc6129e455bf025a5966fc590afe5550e0a0aff06e14d588aa280624"} Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.892670 4815 scope.go:117] "RemoveContainer" containerID="175e366dd5e953a14c204f9dec873e4e240ebdd20e12c52e0d7b74e0d64573cb" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.892957 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.900405 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skz7s" event={"ID":"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd","Type":"ContainerStarted","Data":"138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc"} Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.921553 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/db8bc583-6bcf-40c7-875b-d5c9544af543-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.921790 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.922597 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.923081 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/db8bc583-6bcf-40c7-875b-d5c9544af543-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.924248 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.924409 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/db8bc583-6bcf-40c7-875b-d5c9544af543-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.924522 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db8bc583-6bcf-40c7-875b-d5c9544af543-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.924684 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/db8bc583-6bcf-40c7-875b-d5c9544af543-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.924831 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.924944 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sdqq\" (UniqueName: \"kubernetes.io/projected/db8bc583-6bcf-40c7-875b-d5c9544af543-kube-api-access-9sdqq\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.925038 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.925368 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.924177 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/db8bc583-6bcf-40c7-875b-d5c9544af543-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.923024 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.927270 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.928512 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.929368 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/db8bc583-6bcf-40c7-875b-d5c9544af543-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.922924 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/db8bc583-6bcf-40c7-875b-d5c9544af543-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.933764 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/db8bc583-6bcf-40c7-875b-d5c9544af543-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.944194 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/db8bc583-6bcf-40c7-875b-d5c9544af543-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.945351 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/db8bc583-6bcf-40c7-875b-d5c9544af543-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.949794 4815 scope.go:117] "RemoveContainer" containerID="396633accd6c07faf0270b697296080712b2c885fe98668cd948cdf27a9bfbfa" Dec 05 09:32:32 crc kubenswrapper[4815]: I1205 09:32:32.963503 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sdqq\" (UniqueName: \"kubernetes.io/projected/db8bc583-6bcf-40c7-875b-d5c9544af543-kube-api-access-9sdqq\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.048089 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"db8bc583-6bcf-40c7-875b-d5c9544af543\") " pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.058726 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.062569 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.073094 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.074923 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.078164 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.078297 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.078460 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.078793 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-628cb" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.079705 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.079857 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.080009 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.081966 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.231656 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prj9r\" (UniqueName: \"kubernetes.io/projected/d0d7070e-5d74-4d40-9fe9-b42c60585002-kube-api-access-prj9r\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.231696 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.231724 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.231892 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d0d7070e-5d74-4d40-9fe9-b42c60585002-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.232019 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d0d7070e-5d74-4d40-9fe9-b42c60585002-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.232052 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.232069 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.232120 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.232142 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d0d7070e-5d74-4d40-9fe9-b42c60585002-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.232164 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d0d7070e-5d74-4d40-9fe9-b42c60585002-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.232197 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0d7070e-5d74-4d40-9fe9-b42c60585002-config-data\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.250227 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.333706 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prj9r\" (UniqueName: \"kubernetes.io/projected/d0d7070e-5d74-4d40-9fe9-b42c60585002-kube-api-access-prj9r\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.333751 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.333774 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.333803 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d0d7070e-5d74-4d40-9fe9-b42c60585002-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.333840 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d0d7070e-5d74-4d40-9fe9-b42c60585002-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.333866 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.333911 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.333946 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.333971 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d0d7070e-5d74-4d40-9fe9-b42c60585002-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.333993 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d0d7070e-5d74-4d40-9fe9-b42c60585002-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.334043 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0d7070e-5d74-4d40-9fe9-b42c60585002-config-data\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.334226 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.335031 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.335073 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d0d7070e-5d74-4d40-9fe9-b42c60585002-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.335082 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d0d7070e-5d74-4d40-9fe9-b42c60585002-config-data\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.335259 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.335787 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d0d7070e-5d74-4d40-9fe9-b42c60585002-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.340227 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.342616 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d0d7070e-5d74-4d40-9fe9-b42c60585002-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.347031 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d0d7070e-5d74-4d40-9fe9-b42c60585002-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.348248 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d0d7070e-5d74-4d40-9fe9-b42c60585002-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.355022 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prj9r\" (UniqueName: \"kubernetes.io/projected/d0d7070e-5d74-4d40-9fe9-b42c60585002-kube-api-access-prj9r\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.384124 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"d0d7070e-5d74-4d40-9fe9-b42c60585002\") " pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.403238 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.445348 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="852e4372-ed4f-4350-b013-ab36741cd980" path="/var/lib/kubelet/pods/852e4372-ed4f-4350-b013-ab36741cd980/volumes" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.446225 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e965503d-d60a-4ce1-9dee-b74fd5eb66a0" path="/var/lib/kubelet/pods/e965503d-d60a-4ce1-9dee-b74fd5eb66a0/volumes" Dec 05 09:32:33 crc kubenswrapper[4815]: I1205 09:32:33.995675 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 05 09:32:34 crc kubenswrapper[4815]: W1205 09:32:34.023809 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb8bc583_6bcf_40c7_875b_d5c9544af543.slice/crio-45050f6c45d7adf7233d0f8e85eb38b78e058d5180f6c666ba50766db3479276 WatchSource:0}: Error finding container 45050f6c45d7adf7233d0f8e85eb38b78e058d5180f6c666ba50766db3479276: Status 404 returned error can't find the container with id 45050f6c45d7adf7233d0f8e85eb38b78e058d5180f6c666ba50766db3479276 Dec 05 09:32:34 crc kubenswrapper[4815]: W1205 09:32:34.029056 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0d7070e_5d74_4d40_9fe9_b42c60585002.slice/crio-d7628d3d77ae0ce5b608c42cc28d6c56b07c9c771c787cc0e9b0da00e1dd50f6 WatchSource:0}: Error finding container d7628d3d77ae0ce5b608c42cc28d6c56b07c9c771c787cc0e9b0da00e1dd50f6: Status 404 returned error can't find the container with id d7628d3d77ae0ce5b608c42cc28d6c56b07c9c771c787cc0e9b0da00e1dd50f6 Dec 05 09:32:34 crc kubenswrapper[4815]: I1205 09:32:34.029410 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 05 09:32:34 crc kubenswrapper[4815]: I1205 09:32:34.926983 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"db8bc583-6bcf-40c7-875b-d5c9544af543","Type":"ContainerStarted","Data":"45050f6c45d7adf7233d0f8e85eb38b78e058d5180f6c666ba50766db3479276"} Dec 05 09:32:34 crc kubenswrapper[4815]: I1205 09:32:34.928683 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d0d7070e-5d74-4d40-9fe9-b42c60585002","Type":"ContainerStarted","Data":"d7628d3d77ae0ce5b608c42cc28d6c56b07c9c771c787cc0e9b0da00e1dd50f6"} Dec 05 09:32:34 crc kubenswrapper[4815]: I1205 09:32:34.932418 4815 generic.go:334] "Generic (PLEG): container finished" podID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" containerID="138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc" exitCode=0 Dec 05 09:32:34 crc kubenswrapper[4815]: I1205 09:32:34.932720 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skz7s" event={"ID":"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd","Type":"ContainerDied","Data":"138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc"} Dec 05 09:32:35 crc kubenswrapper[4815]: I1205 09:32:35.379406 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:35 crc kubenswrapper[4815]: I1205 09:32:35.379484 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:35 crc kubenswrapper[4815]: I1205 09:32:35.439004 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:35 crc kubenswrapper[4815]: I1205 09:32:35.942543 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skz7s" event={"ID":"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd","Type":"ContainerStarted","Data":"aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc"} Dec 05 09:32:35 crc kubenswrapper[4815]: I1205 09:32:35.946851 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"db8bc583-6bcf-40c7-875b-d5c9544af543","Type":"ContainerStarted","Data":"bef425d61df5a105484263d59b2791a968ce33696f697df94f955f803acf06d3"} Dec 05 09:32:35 crc kubenswrapper[4815]: I1205 09:32:35.949227 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d0d7070e-5d74-4d40-9fe9-b42c60585002","Type":"ContainerStarted","Data":"17e35b6e1511ba1cd9ab015afe69b52527d1cf65973f4fce03b0c4ba66b48593"} Dec 05 09:32:35 crc kubenswrapper[4815]: I1205 09:32:35.964213 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-skz7s" podStartSLOduration=2.318885103 podStartE2EDuration="6.964194316s" podCreationTimestamp="2025-12-05 09:32:29 +0000 UTC" firstStartedPulling="2025-12-05 09:32:30.781090987 +0000 UTC m=+1549.659697824" lastFinishedPulling="2025-12-05 09:32:35.42640018 +0000 UTC m=+1554.305007037" observedRunningTime="2025-12-05 09:32:35.963061516 +0000 UTC m=+1554.841668353" watchObservedRunningTime="2025-12-05 09:32:35.964194316 +0000 UTC m=+1554.842801153" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.249124 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8676544467-85bq4"] Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.251104 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.253515 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.269163 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8676544467-85bq4"] Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.285996 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-sb\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.286104 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-nb\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.286167 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-config\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.286263 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-dns-svc\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.286292 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gkvs4\" (UniqueName: \"kubernetes.io/projected/f57b7d3c-3e12-4a19-b467-4c0566d51b99-kube-api-access-gkvs4\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.286321 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-openstack-edpm-ipam\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.388105 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-sb\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.388182 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-nb\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.388211 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-config\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.388270 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-dns-svc\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.388289 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gkvs4\" (UniqueName: \"kubernetes.io/projected/f57b7d3c-3e12-4a19-b467-4c0566d51b99-kube-api-access-gkvs4\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.388311 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-openstack-edpm-ipam\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.389037 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-sb\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.389101 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-openstack-edpm-ipam\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.389144 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-nb\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.389690 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-config\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.390402 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-dns-svc\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.408868 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gkvs4\" (UniqueName: \"kubernetes.io/projected/f57b7d3c-3e12-4a19-b467-4c0566d51b99-kube-api-access-gkvs4\") pod \"dnsmasq-dns-8676544467-85bq4\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.570074 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.686789 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.688182 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:39 crc kubenswrapper[4815]: I1205 09:32:39.774074 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:40 crc kubenswrapper[4815]: I1205 09:32:40.175071 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:40 crc kubenswrapper[4815]: I1205 09:32:40.191020 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8676544467-85bq4"] Dec 05 09:32:40 crc kubenswrapper[4815]: W1205 09:32:40.208128 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf57b7d3c_3e12_4a19_b467_4c0566d51b99.slice/crio-08c6482971ad3869890e813f6d2012897c047c2d9d8cc389312abdf0c742d7ed WatchSource:0}: Error finding container 08c6482971ad3869890e813f6d2012897c047c2d9d8cc389312abdf0c742d7ed: Status 404 returned error can't find the container with id 08c6482971ad3869890e813f6d2012897c047c2d9d8cc389312abdf0c742d7ed Dec 05 09:32:40 crc kubenswrapper[4815]: I1205 09:32:40.343543 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-skz7s"] Dec 05 09:32:41 crc kubenswrapper[4815]: I1205 09:32:41.012833 4815 generic.go:334] "Generic (PLEG): container finished" podID="f57b7d3c-3e12-4a19-b467-4c0566d51b99" containerID="c217725269bb550fc9c48f75bd4e521476d25b953d998c82c1404b68e8642e12" exitCode=0 Dec 05 09:32:41 crc kubenswrapper[4815]: I1205 09:32:41.012875 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8676544467-85bq4" event={"ID":"f57b7d3c-3e12-4a19-b467-4c0566d51b99","Type":"ContainerDied","Data":"c217725269bb550fc9c48f75bd4e521476d25b953d998c82c1404b68e8642e12"} Dec 05 09:32:41 crc kubenswrapper[4815]: I1205 09:32:41.013158 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8676544467-85bq4" event={"ID":"f57b7d3c-3e12-4a19-b467-4c0566d51b99","Type":"ContainerStarted","Data":"08c6482971ad3869890e813f6d2012897c047c2d9d8cc389312abdf0c742d7ed"} Dec 05 09:32:42 crc kubenswrapper[4815]: I1205 09:32:42.025177 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-skz7s" podUID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" containerName="registry-server" containerID="cri-o://aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc" gracePeriod=2 Dec 05 09:32:42 crc kubenswrapper[4815]: I1205 09:32:42.025517 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8676544467-85bq4" event={"ID":"f57b7d3c-3e12-4a19-b467-4c0566d51b99","Type":"ContainerStarted","Data":"f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56"} Dec 05 09:32:42 crc kubenswrapper[4815]: I1205 09:32:42.025984 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:42 crc kubenswrapper[4815]: I1205 09:32:42.063205 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8676544467-85bq4" podStartSLOduration=3.063182041 podStartE2EDuration="3.063182041s" podCreationTimestamp="2025-12-05 09:32:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:32:42.057120923 +0000 UTC m=+1560.935727790" watchObservedRunningTime="2025-12-05 09:32:42.063182041 +0000 UTC m=+1560.941788898" Dec 05 09:32:42 crc kubenswrapper[4815]: I1205 09:32:42.738175 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:42 crc kubenswrapper[4815]: I1205 09:32:42.917725 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-catalog-content\") pod \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " Dec 05 09:32:42 crc kubenswrapper[4815]: I1205 09:32:42.917799 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5l9n\" (UniqueName: \"kubernetes.io/projected/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-kube-api-access-f5l9n\") pod \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " Dec 05 09:32:42 crc kubenswrapper[4815]: I1205 09:32:42.918534 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-utilities\") pod \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\" (UID: \"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd\") " Dec 05 09:32:42 crc kubenswrapper[4815]: I1205 09:32:42.919164 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-utilities" (OuterVolumeSpecName: "utilities") pod "e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" (UID: "e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:32:42 crc kubenswrapper[4815]: I1205 09:32:42.924376 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-kube-api-access-f5l9n" (OuterVolumeSpecName: "kube-api-access-f5l9n") pod "e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" (UID: "e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd"). InnerVolumeSpecName "kube-api-access-f5l9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:32:42 crc kubenswrapper[4815]: I1205 09:32:42.959371 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" (UID: "e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.021055 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.021092 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.021110 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5l9n\" (UniqueName: \"kubernetes.io/projected/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd-kube-api-access-f5l9n\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.036742 4815 generic.go:334] "Generic (PLEG): container finished" podID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" containerID="aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc" exitCode=0 Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.037741 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-skz7s" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.040757 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skz7s" event={"ID":"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd","Type":"ContainerDied","Data":"aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc"} Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.040822 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-skz7s" event={"ID":"e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd","Type":"ContainerDied","Data":"c43633b2e8a81172a41542caaa3e8a542b2de761b5b628ccfea439f2ae75f3b6"} Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.040840 4815 scope.go:117] "RemoveContainer" containerID="aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.075387 4815 scope.go:117] "RemoveContainer" containerID="138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.093769 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-skz7s"] Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.110470 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-skz7s"] Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.124532 4815 scope.go:117] "RemoveContainer" containerID="d0ad33a617d68fdf4b0c4062d5a7b8145ad83fb3a8addbeb0bbdbc3087ec3d82" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.153810 4815 scope.go:117] "RemoveContainer" containerID="aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc" Dec 05 09:32:43 crc kubenswrapper[4815]: E1205 09:32:43.154731 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc\": container with ID starting with aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc not found: ID does not exist" containerID="aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.154765 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc"} err="failed to get container status \"aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc\": rpc error: code = NotFound desc = could not find container \"aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc\": container with ID starting with aac425e45de5d06c67d507f9235d571134049231efe3d3f82b94b0f1b403b9dc not found: ID does not exist" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.154790 4815 scope.go:117] "RemoveContainer" containerID="138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc" Dec 05 09:32:43 crc kubenswrapper[4815]: E1205 09:32:43.155086 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc\": container with ID starting with 138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc not found: ID does not exist" containerID="138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.155134 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc"} err="failed to get container status \"138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc\": rpc error: code = NotFound desc = could not find container \"138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc\": container with ID starting with 138e8c39fadf952a3f94d98c2b1020c455906fd96908715ed4e02b80c7a4f4bc not found: ID does not exist" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.155148 4815 scope.go:117] "RemoveContainer" containerID="d0ad33a617d68fdf4b0c4062d5a7b8145ad83fb3a8addbeb0bbdbc3087ec3d82" Dec 05 09:32:43 crc kubenswrapper[4815]: E1205 09:32:43.155535 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0ad33a617d68fdf4b0c4062d5a7b8145ad83fb3a8addbeb0bbdbc3087ec3d82\": container with ID starting with d0ad33a617d68fdf4b0c4062d5a7b8145ad83fb3a8addbeb0bbdbc3087ec3d82 not found: ID does not exist" containerID="d0ad33a617d68fdf4b0c4062d5a7b8145ad83fb3a8addbeb0bbdbc3087ec3d82" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.155557 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0ad33a617d68fdf4b0c4062d5a7b8145ad83fb3a8addbeb0bbdbc3087ec3d82"} err="failed to get container status \"d0ad33a617d68fdf4b0c4062d5a7b8145ad83fb3a8addbeb0bbdbc3087ec3d82\": rpc error: code = NotFound desc = could not find container \"d0ad33a617d68fdf4b0c4062d5a7b8145ad83fb3a8addbeb0bbdbc3087ec3d82\": container with ID starting with d0ad33a617d68fdf4b0c4062d5a7b8145ad83fb3a8addbeb0bbdbc3087ec3d82 not found: ID does not exist" Dec 05 09:32:43 crc kubenswrapper[4815]: I1205 09:32:43.430222 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" path="/var/lib/kubelet/pods/e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd/volumes" Dec 05 09:32:45 crc kubenswrapper[4815]: I1205 09:32:45.436473 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:45 crc kubenswrapper[4815]: I1205 09:32:45.488527 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdkcx"] Dec 05 09:32:46 crc kubenswrapper[4815]: I1205 09:32:46.075673 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qdkcx" podUID="ce5b001e-1094-426f-b412-6d6042c3e526" containerName="registry-server" containerID="cri-o://c39db0357a6e90c90581575b6431bc8843876a1bb495f8bd5939eeab8e8976eb" gracePeriod=2 Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.102433 4815 generic.go:334] "Generic (PLEG): container finished" podID="ce5b001e-1094-426f-b412-6d6042c3e526" containerID="c39db0357a6e90c90581575b6431bc8843876a1bb495f8bd5939eeab8e8976eb" exitCode=0 Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.102817 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdkcx" event={"ID":"ce5b001e-1094-426f-b412-6d6042c3e526","Type":"ContainerDied","Data":"c39db0357a6e90c90581575b6431bc8843876a1bb495f8bd5939eeab8e8976eb"} Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.292789 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.441992 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-utilities\") pod \"ce5b001e-1094-426f-b412-6d6042c3e526\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.442446 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-catalog-content\") pod \"ce5b001e-1094-426f-b412-6d6042c3e526\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.442987 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-utilities" (OuterVolumeSpecName: "utilities") pod "ce5b001e-1094-426f-b412-6d6042c3e526" (UID: "ce5b001e-1094-426f-b412-6d6042c3e526"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.443392 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fh5ff\" (UniqueName: \"kubernetes.io/projected/ce5b001e-1094-426f-b412-6d6042c3e526-kube-api-access-fh5ff\") pod \"ce5b001e-1094-426f-b412-6d6042c3e526\" (UID: \"ce5b001e-1094-426f-b412-6d6042c3e526\") " Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.444436 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.451385 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce5b001e-1094-426f-b412-6d6042c3e526-kube-api-access-fh5ff" (OuterVolumeSpecName: "kube-api-access-fh5ff") pod "ce5b001e-1094-426f-b412-6d6042c3e526" (UID: "ce5b001e-1094-426f-b412-6d6042c3e526"). InnerVolumeSpecName "kube-api-access-fh5ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.467206 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce5b001e-1094-426f-b412-6d6042c3e526" (UID: "ce5b001e-1094-426f-b412-6d6042c3e526"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.548874 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fh5ff\" (UniqueName: \"kubernetes.io/projected/ce5b001e-1094-426f-b412-6d6042c3e526-kube-api-access-fh5ff\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:47 crc kubenswrapper[4815]: I1205 09:32:47.548913 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce5b001e-1094-426f-b412-6d6042c3e526-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:48 crc kubenswrapper[4815]: I1205 09:32:48.113729 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qdkcx" event={"ID":"ce5b001e-1094-426f-b412-6d6042c3e526","Type":"ContainerDied","Data":"be0bf9b35f5cd0385ac758b6409e461731a2adbdeb2324a2bd30c38562c68175"} Dec 05 09:32:48 crc kubenswrapper[4815]: I1205 09:32:48.114000 4815 scope.go:117] "RemoveContainer" containerID="c39db0357a6e90c90581575b6431bc8843876a1bb495f8bd5939eeab8e8976eb" Dec 05 09:32:48 crc kubenswrapper[4815]: I1205 09:32:48.113813 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qdkcx" Dec 05 09:32:48 crc kubenswrapper[4815]: I1205 09:32:48.141634 4815 scope.go:117] "RemoveContainer" containerID="d76a6a7ffea0c39b2d609264cd5957d06a26bc215956efa5fc0fc18436fa9ad7" Dec 05 09:32:48 crc kubenswrapper[4815]: I1205 09:32:48.156636 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdkcx"] Dec 05 09:32:48 crc kubenswrapper[4815]: I1205 09:32:48.171773 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qdkcx"] Dec 05 09:32:48 crc kubenswrapper[4815]: I1205 09:32:48.187617 4815 scope.go:117] "RemoveContainer" containerID="3ce7dd083e49a3124ef884ae63672121e67c4117bfd647d9728521279427e824" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.437122 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce5b001e-1094-426f-b412-6d6042c3e526" path="/var/lib/kubelet/pods/ce5b001e-1094-426f-b412-6d6042c3e526/volumes" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.571645 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.656879 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f64966d5-qbnzb"] Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.657215 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" podUID="d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" containerName="dnsmasq-dns" containerID="cri-o://2e1d416e21594c96355f5509734f19c8ba57cde8b545bba1124a2d3bf5ee8131" gracePeriod=10 Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.862628 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-54fcc757dc-kbs2f"] Dec 05 09:32:49 crc kubenswrapper[4815]: E1205 09:32:49.863003 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5b001e-1094-426f-b412-6d6042c3e526" containerName="extract-utilities" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.863027 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5b001e-1094-426f-b412-6d6042c3e526" containerName="extract-utilities" Dec 05 09:32:49 crc kubenswrapper[4815]: E1205 09:32:49.863044 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" containerName="registry-server" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.863051 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" containerName="registry-server" Dec 05 09:32:49 crc kubenswrapper[4815]: E1205 09:32:49.863068 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5b001e-1094-426f-b412-6d6042c3e526" containerName="registry-server" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.863075 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5b001e-1094-426f-b412-6d6042c3e526" containerName="registry-server" Dec 05 09:32:49 crc kubenswrapper[4815]: E1205 09:32:49.863090 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce5b001e-1094-426f-b412-6d6042c3e526" containerName="extract-content" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.863095 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce5b001e-1094-426f-b412-6d6042c3e526" containerName="extract-content" Dec 05 09:32:49 crc kubenswrapper[4815]: E1205 09:32:49.863107 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" containerName="extract-utilities" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.863112 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" containerName="extract-utilities" Dec 05 09:32:49 crc kubenswrapper[4815]: E1205 09:32:49.863124 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" containerName="extract-content" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.863129 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" containerName="extract-content" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.863298 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7edeaa8-e3bd-4535-bf1f-1d08268ea0dd" containerName="registry-server" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.863311 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce5b001e-1094-426f-b412-6d6042c3e526" containerName="registry-server" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.864194 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.889170 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54fcc757dc-kbs2f"] Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.983707 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-openstack-edpm-ipam\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.984026 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-sb\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.984058 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-dns-svc\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.984084 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-nb\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.984121 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m222v\" (UniqueName: \"kubernetes.io/projected/a06a9288-ec09-4ff0-8a21-c0a131880da8-kube-api-access-m222v\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:49 crc kubenswrapper[4815]: I1205 09:32:49.984221 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-config\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.086054 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-config\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.086182 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-openstack-edpm-ipam\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.086218 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-sb\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.086255 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-dns-svc\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.086294 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-nb\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.086344 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m222v\" (UniqueName: \"kubernetes.io/projected/a06a9288-ec09-4ff0-8a21-c0a131880da8-kube-api-access-m222v\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.087097 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-openstack-edpm-ipam\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.087629 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-sb\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.087798 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-dns-svc\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.088340 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-nb\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.088555 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-config\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.137902 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m222v\" (UniqueName: \"kubernetes.io/projected/a06a9288-ec09-4ff0-8a21-c0a131880da8-kube-api-access-m222v\") pod \"dnsmasq-dns-54fcc757dc-kbs2f\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.150397 4815 generic.go:334] "Generic (PLEG): container finished" podID="d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" containerID="2e1d416e21594c96355f5509734f19c8ba57cde8b545bba1124a2d3bf5ee8131" exitCode=0 Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.150442 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" event={"ID":"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7","Type":"ContainerDied","Data":"2e1d416e21594c96355f5509734f19c8ba57cde8b545bba1124a2d3bf5ee8131"} Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.181435 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.192255 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.192607 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.324675 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.501384 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-sb\") pod \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.501472 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-nb\") pod \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.501615 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpmw2\" (UniqueName: \"kubernetes.io/projected/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-kube-api-access-qpmw2\") pod \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.501658 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-config\") pod \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.501754 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-dns-svc\") pod \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\" (UID: \"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7\") " Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.512259 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-kube-api-access-qpmw2" (OuterVolumeSpecName: "kube-api-access-qpmw2") pod "d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" (UID: "d577dab3-e67f-4bc3-b96c-55e40eb7c9a7"). InnerVolumeSpecName "kube-api-access-qpmw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.566074 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" (UID: "d577dab3-e67f-4bc3-b96c-55e40eb7c9a7"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.581861 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" (UID: "d577dab3-e67f-4bc3-b96c-55e40eb7c9a7"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.590084 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" (UID: "d577dab3-e67f-4bc3-b96c-55e40eb7c9a7"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.604301 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpmw2\" (UniqueName: \"kubernetes.io/projected/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-kube-api-access-qpmw2\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.604334 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.604345 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.604353 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.618113 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-config" (OuterVolumeSpecName: "config") pod "d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" (UID: "d577dab3-e67f-4bc3-b96c-55e40eb7c9a7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.685831 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-54fcc757dc-kbs2f"] Dec 05 09:32:50 crc kubenswrapper[4815]: I1205 09:32:50.707216 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:32:51 crc kubenswrapper[4815]: I1205 09:32:51.160031 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" event={"ID":"d577dab3-e67f-4bc3-b96c-55e40eb7c9a7","Type":"ContainerDied","Data":"6f1e32e6749c6efb54fdf481e947c39908cc0642166744d529058860ebe98ed3"} Dec 05 09:32:51 crc kubenswrapper[4815]: I1205 09:32:51.160369 4815 scope.go:117] "RemoveContainer" containerID="2e1d416e21594c96355f5509734f19c8ba57cde8b545bba1124a2d3bf5ee8131" Dec 05 09:32:51 crc kubenswrapper[4815]: I1205 09:32:51.160211 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f64966d5-qbnzb" Dec 05 09:32:51 crc kubenswrapper[4815]: I1205 09:32:51.162611 4815 generic.go:334] "Generic (PLEG): container finished" podID="a06a9288-ec09-4ff0-8a21-c0a131880da8" containerID="28f263330a6907540525a3495bdb79ddbb3e0ea4b51331dd2dfc415185e384c2" exitCode=0 Dec 05 09:32:51 crc kubenswrapper[4815]: I1205 09:32:51.162714 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" event={"ID":"a06a9288-ec09-4ff0-8a21-c0a131880da8","Type":"ContainerDied","Data":"28f263330a6907540525a3495bdb79ddbb3e0ea4b51331dd2dfc415185e384c2"} Dec 05 09:32:51 crc kubenswrapper[4815]: I1205 09:32:51.162808 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" event={"ID":"a06a9288-ec09-4ff0-8a21-c0a131880da8","Type":"ContainerStarted","Data":"233da146f2f2de53321c030bb98a228cc959c0582eda3cab3a1b376175f58658"} Dec 05 09:32:51 crc kubenswrapper[4815]: I1205 09:32:51.178417 4815 scope.go:117] "RemoveContainer" containerID="00dcefd610fc0d38f025a679f158ac903e490264b824ef7b651fd4e2531f6ae6" Dec 05 09:32:51 crc kubenswrapper[4815]: I1205 09:32:51.274866 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f64966d5-qbnzb"] Dec 05 09:32:51 crc kubenswrapper[4815]: I1205 09:32:51.284599 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f64966d5-qbnzb"] Dec 05 09:32:51 crc kubenswrapper[4815]: I1205 09:32:51.431293 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" path="/var/lib/kubelet/pods/d577dab3-e67f-4bc3-b96c-55e40eb7c9a7/volumes" Dec 05 09:32:52 crc kubenswrapper[4815]: I1205 09:32:52.180327 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" event={"ID":"a06a9288-ec09-4ff0-8a21-c0a131880da8","Type":"ContainerStarted","Data":"ee9878b701d75509d14856c88dd79be3eef3919f11c76c7f390446563bd86c7e"} Dec 05 09:32:52 crc kubenswrapper[4815]: I1205 09:32:52.180812 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:32:52 crc kubenswrapper[4815]: I1205 09:32:52.214948 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" podStartSLOduration=3.214929626 podStartE2EDuration="3.214929626s" podCreationTimestamp="2025-12-05 09:32:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:32:52.20423425 +0000 UTC m=+1571.082841087" watchObservedRunningTime="2025-12-05 09:32:52.214929626 +0000 UTC m=+1571.093536463" Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.183685 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.258110 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8676544467-85bq4"] Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.258381 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8676544467-85bq4" podUID="f57b7d3c-3e12-4a19-b467-4c0566d51b99" containerName="dnsmasq-dns" containerID="cri-o://f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56" gracePeriod=10 Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.770184 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.814388 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-openstack-edpm-ipam\") pod \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.883049 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "f57b7d3c-3e12-4a19-b467-4c0566d51b99" (UID: "f57b7d3c-3e12-4a19-b467-4c0566d51b99"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.915522 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-sb\") pod \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.915566 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-config\") pod \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.915637 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-nb\") pod \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.915662 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gkvs4\" (UniqueName: \"kubernetes.io/projected/f57b7d3c-3e12-4a19-b467-4c0566d51b99-kube-api-access-gkvs4\") pod \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.915690 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-dns-svc\") pod \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\" (UID: \"f57b7d3c-3e12-4a19-b467-4c0566d51b99\") " Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.915979 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.921201 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f57b7d3c-3e12-4a19-b467-4c0566d51b99-kube-api-access-gkvs4" (OuterVolumeSpecName: "kube-api-access-gkvs4") pod "f57b7d3c-3e12-4a19-b467-4c0566d51b99" (UID: "f57b7d3c-3e12-4a19-b467-4c0566d51b99"). InnerVolumeSpecName "kube-api-access-gkvs4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.974005 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f57b7d3c-3e12-4a19-b467-4c0566d51b99" (UID: "f57b7d3c-3e12-4a19-b467-4c0566d51b99"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.983271 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f57b7d3c-3e12-4a19-b467-4c0566d51b99" (UID: "f57b7d3c-3e12-4a19-b467-4c0566d51b99"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:33:00 crc kubenswrapper[4815]: I1205 09:33:00.986677 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-config" (OuterVolumeSpecName: "config") pod "f57b7d3c-3e12-4a19-b467-4c0566d51b99" (UID: "f57b7d3c-3e12-4a19-b467-4c0566d51b99"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.004097 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f57b7d3c-3e12-4a19-b467-4c0566d51b99" (UID: "f57b7d3c-3e12-4a19-b467-4c0566d51b99"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.017160 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.017385 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-config\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.017446 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.017521 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gkvs4\" (UniqueName: \"kubernetes.io/projected/f57b7d3c-3e12-4a19-b467-4c0566d51b99-kube-api-access-gkvs4\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.017580 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f57b7d3c-3e12-4a19-b467-4c0566d51b99-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.261609 4815 generic.go:334] "Generic (PLEG): container finished" podID="f57b7d3c-3e12-4a19-b467-4c0566d51b99" containerID="f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56" exitCode=0 Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.261652 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8676544467-85bq4" event={"ID":"f57b7d3c-3e12-4a19-b467-4c0566d51b99","Type":"ContainerDied","Data":"f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56"} Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.261682 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8676544467-85bq4" event={"ID":"f57b7d3c-3e12-4a19-b467-4c0566d51b99","Type":"ContainerDied","Data":"08c6482971ad3869890e813f6d2012897c047c2d9d8cc389312abdf0c742d7ed"} Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.261689 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8676544467-85bq4" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.261704 4815 scope.go:117] "RemoveContainer" containerID="f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.300109 4815 scope.go:117] "RemoveContainer" containerID="c217725269bb550fc9c48f75bd4e521476d25b953d998c82c1404b68e8642e12" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.317152 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8676544467-85bq4"] Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.320638 4815 scope.go:117] "RemoveContainer" containerID="f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56" Dec 05 09:33:01 crc kubenswrapper[4815]: E1205 09:33:01.321957 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56\": container with ID starting with f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56 not found: ID does not exist" containerID="f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.322017 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56"} err="failed to get container status \"f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56\": rpc error: code = NotFound desc = could not find container \"f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56\": container with ID starting with f8d6a77bf6bef64324151d89337f344acfcc353658ab3dacb66b56737da9ab56 not found: ID does not exist" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.322054 4815 scope.go:117] "RemoveContainer" containerID="c217725269bb550fc9c48f75bd4e521476d25b953d998c82c1404b68e8642e12" Dec 05 09:33:01 crc kubenswrapper[4815]: E1205 09:33:01.322323 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c217725269bb550fc9c48f75bd4e521476d25b953d998c82c1404b68e8642e12\": container with ID starting with c217725269bb550fc9c48f75bd4e521476d25b953d998c82c1404b68e8642e12 not found: ID does not exist" containerID="c217725269bb550fc9c48f75bd4e521476d25b953d998c82c1404b68e8642e12" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.322350 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c217725269bb550fc9c48f75bd4e521476d25b953d998c82c1404b68e8642e12"} err="failed to get container status \"c217725269bb550fc9c48f75bd4e521476d25b953d998c82c1404b68e8642e12\": rpc error: code = NotFound desc = could not find container \"c217725269bb550fc9c48f75bd4e521476d25b953d998c82c1404b68e8642e12\": container with ID starting with c217725269bb550fc9c48f75bd4e521476d25b953d998c82c1404b68e8642e12 not found: ID does not exist" Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.327467 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8676544467-85bq4"] Dec 05 09:33:01 crc kubenswrapper[4815]: I1205 09:33:01.434399 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f57b7d3c-3e12-4a19-b467-4c0566d51b99" path="/var/lib/kubelet/pods/f57b7d3c-3e12-4a19-b467-4c0566d51b99/volumes" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.337648 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k"] Dec 05 09:33:06 crc kubenswrapper[4815]: E1205 09:33:06.338991 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" containerName="init" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.339172 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" containerName="init" Dec 05 09:33:06 crc kubenswrapper[4815]: E1205 09:33:06.339200 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f57b7d3c-3e12-4a19-b467-4c0566d51b99" containerName="init" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.339208 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f57b7d3c-3e12-4a19-b467-4c0566d51b99" containerName="init" Dec 05 09:33:06 crc kubenswrapper[4815]: E1205 09:33:06.339265 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" containerName="dnsmasq-dns" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.339273 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" containerName="dnsmasq-dns" Dec 05 09:33:06 crc kubenswrapper[4815]: E1205 09:33:06.339301 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f57b7d3c-3e12-4a19-b467-4c0566d51b99" containerName="dnsmasq-dns" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.339314 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f57b7d3c-3e12-4a19-b467-4c0566d51b99" containerName="dnsmasq-dns" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.339772 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d577dab3-e67f-4bc3-b96c-55e40eb7c9a7" containerName="dnsmasq-dns" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.339801 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f57b7d3c-3e12-4a19-b467-4c0566d51b99" containerName="dnsmasq-dns" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.341053 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.351700 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.351780 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.351929 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.352062 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.369251 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k"] Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.418174 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.418454 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.418600 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggqh7\" (UniqueName: \"kubernetes.io/projected/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-kube-api-access-ggqh7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.418746 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.520551 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.520638 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggqh7\" (UniqueName: \"kubernetes.io/projected/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-kube-api-access-ggqh7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.520676 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.520746 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.527177 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.540277 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.540891 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.544282 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggqh7\" (UniqueName: \"kubernetes.io/projected/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-kube-api-access-ggqh7\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:06 crc kubenswrapper[4815]: I1205 09:33:06.677816 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:07 crc kubenswrapper[4815]: I1205 09:33:07.894792 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k"] Dec 05 09:33:07 crc kubenswrapper[4815]: W1205 09:33:07.896380 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2212e6a0_bdfa_41e1_8ff3_2b386ff23864.slice/crio-44b8cb718468d0ab317fdd8c61f0868f55d5598b8d1656f9bb1c12166e7b1618 WatchSource:0}: Error finding container 44b8cb718468d0ab317fdd8c61f0868f55d5598b8d1656f9bb1c12166e7b1618: Status 404 returned error can't find the container with id 44b8cb718468d0ab317fdd8c61f0868f55d5598b8d1656f9bb1c12166e7b1618 Dec 05 09:33:08 crc kubenswrapper[4815]: I1205 09:33:08.337717 4815 generic.go:334] "Generic (PLEG): container finished" podID="db8bc583-6bcf-40c7-875b-d5c9544af543" containerID="bef425d61df5a105484263d59b2791a968ce33696f697df94f955f803acf06d3" exitCode=0 Dec 05 09:33:08 crc kubenswrapper[4815]: I1205 09:33:08.337753 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"db8bc583-6bcf-40c7-875b-d5c9544af543","Type":"ContainerDied","Data":"bef425d61df5a105484263d59b2791a968ce33696f697df94f955f803acf06d3"} Dec 05 09:33:08 crc kubenswrapper[4815]: I1205 09:33:08.339677 4815 generic.go:334] "Generic (PLEG): container finished" podID="d0d7070e-5d74-4d40-9fe9-b42c60585002" containerID="17e35b6e1511ba1cd9ab015afe69b52527d1cf65973f4fce03b0c4ba66b48593" exitCode=0 Dec 05 09:33:08 crc kubenswrapper[4815]: I1205 09:33:08.339769 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d0d7070e-5d74-4d40-9fe9-b42c60585002","Type":"ContainerDied","Data":"17e35b6e1511ba1cd9ab015afe69b52527d1cf65973f4fce03b0c4ba66b48593"} Dec 05 09:33:08 crc kubenswrapper[4815]: I1205 09:33:08.341707 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" event={"ID":"2212e6a0-bdfa-41e1-8ff3-2b386ff23864","Type":"ContainerStarted","Data":"44b8cb718468d0ab317fdd8c61f0868f55d5598b8d1656f9bb1c12166e7b1618"} Dec 05 09:33:09 crc kubenswrapper[4815]: I1205 09:33:09.364421 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"db8bc583-6bcf-40c7-875b-d5c9544af543","Type":"ContainerStarted","Data":"2da03bbea6328e5ca085624ee6ad7759fd7775c492ee4342a24792eff9612a61"} Dec 05 09:33:09 crc kubenswrapper[4815]: I1205 09:33:09.366147 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:33:09 crc kubenswrapper[4815]: I1205 09:33:09.373684 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d0d7070e-5d74-4d40-9fe9-b42c60585002","Type":"ContainerStarted","Data":"a4347a91f0cf2b9ece9ff9a819131b7933af9b584eefa1b77f0f76f333337fb8"} Dec 05 09:33:09 crc kubenswrapper[4815]: I1205 09:33:09.374649 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 05 09:33:09 crc kubenswrapper[4815]: I1205 09:33:09.409285 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.40926149 podStartE2EDuration="37.40926149s" podCreationTimestamp="2025-12-05 09:32:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:33:09.40308462 +0000 UTC m=+1588.281691467" watchObservedRunningTime="2025-12-05 09:33:09.40926149 +0000 UTC m=+1588.287868327" Dec 05 09:33:09 crc kubenswrapper[4815]: I1205 09:33:09.448704 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.448679751 podStartE2EDuration="37.448679751s" podCreationTimestamp="2025-12-05 09:32:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 09:33:09.442346197 +0000 UTC m=+1588.320953054" watchObservedRunningTime="2025-12-05 09:33:09.448679751 +0000 UTC m=+1588.327286588" Dec 05 09:33:20 crc kubenswrapper[4815]: I1205 09:33:20.192893 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:33:20 crc kubenswrapper[4815]: I1205 09:33:20.193433 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:33:20 crc kubenswrapper[4815]: I1205 09:33:20.193477 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:33:20 crc kubenswrapper[4815]: I1205 09:33:20.194206 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:33:20 crc kubenswrapper[4815]: I1205 09:33:20.194252 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" gracePeriod=600 Dec 05 09:33:20 crc kubenswrapper[4815]: I1205 09:33:20.490869 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" exitCode=0 Dec 05 09:33:20 crc kubenswrapper[4815]: I1205 09:33:20.490929 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa"} Dec 05 09:33:20 crc kubenswrapper[4815]: I1205 09:33:20.490981 4815 scope.go:117] "RemoveContainer" containerID="ca66886398a56fe8786c896f7db80c714344ffc63db0667dbec913c482105f33" Dec 05 09:33:22 crc kubenswrapper[4815]: E1205 09:33:22.518262 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:33:23 crc kubenswrapper[4815]: I1205 09:33:23.253097 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="db8bc583-6bcf-40c7-875b-d5c9544af543" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.191:5671: connect: connection refused" Dec 05 09:33:23 crc kubenswrapper[4815]: I1205 09:33:23.405111 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="d0d7070e-5d74-4d40-9fe9-b42c60585002" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.192:5671: connect: connection refused" Dec 05 09:33:23 crc kubenswrapper[4815]: I1205 09:33:23.527156 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" event={"ID":"2212e6a0-bdfa-41e1-8ff3-2b386ff23864","Type":"ContainerStarted","Data":"46443ce923ce64a901168a9fc589b424e164362a99a53922866bb1d7c815e1d5"} Dec 05 09:33:23 crc kubenswrapper[4815]: I1205 09:33:23.529575 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:33:23 crc kubenswrapper[4815]: E1205 09:33:23.529803 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:33:23 crc kubenswrapper[4815]: I1205 09:33:23.728620 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" podStartSLOduration=3.00333698 podStartE2EDuration="17.728600702s" podCreationTimestamp="2025-12-05 09:33:06 +0000 UTC" firstStartedPulling="2025-12-05 09:33:07.899828985 +0000 UTC m=+1586.778435822" lastFinishedPulling="2025-12-05 09:33:22.625092707 +0000 UTC m=+1601.503699544" observedRunningTime="2025-12-05 09:33:23.723759877 +0000 UTC m=+1602.602366714" watchObservedRunningTime="2025-12-05 09:33:23.728600702 +0000 UTC m=+1602.607207529" Dec 05 09:33:33 crc kubenswrapper[4815]: I1205 09:33:33.252482 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 05 09:33:33 crc kubenswrapper[4815]: I1205 09:33:33.404874 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 05 09:33:39 crc kubenswrapper[4815]: I1205 09:33:39.418929 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:33:39 crc kubenswrapper[4815]: E1205 09:33:39.419761 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:33:39 crc kubenswrapper[4815]: I1205 09:33:39.672139 4815 generic.go:334] "Generic (PLEG): container finished" podID="2212e6a0-bdfa-41e1-8ff3-2b386ff23864" containerID="46443ce923ce64a901168a9fc589b424e164362a99a53922866bb1d7c815e1d5" exitCode=0 Dec 05 09:33:39 crc kubenswrapper[4815]: I1205 09:33:39.672186 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" event={"ID":"2212e6a0-bdfa-41e1-8ff3-2b386ff23864","Type":"ContainerDied","Data":"46443ce923ce64a901168a9fc589b424e164362a99a53922866bb1d7c815e1d5"} Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.068366 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.135406 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggqh7\" (UniqueName: \"kubernetes.io/projected/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-kube-api-access-ggqh7\") pod \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.135562 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-ssh-key\") pod \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.135599 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-inventory\") pod \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.135723 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-repo-setup-combined-ca-bundle\") pod \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\" (UID: \"2212e6a0-bdfa-41e1-8ff3-2b386ff23864\") " Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.142051 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "2212e6a0-bdfa-41e1-8ff3-2b386ff23864" (UID: "2212e6a0-bdfa-41e1-8ff3-2b386ff23864"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.142800 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-kube-api-access-ggqh7" (OuterVolumeSpecName: "kube-api-access-ggqh7") pod "2212e6a0-bdfa-41e1-8ff3-2b386ff23864" (UID: "2212e6a0-bdfa-41e1-8ff3-2b386ff23864"). InnerVolumeSpecName "kube-api-access-ggqh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.164580 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-inventory" (OuterVolumeSpecName: "inventory") pod "2212e6a0-bdfa-41e1-8ff3-2b386ff23864" (UID: "2212e6a0-bdfa-41e1-8ff3-2b386ff23864"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.168279 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2212e6a0-bdfa-41e1-8ff3-2b386ff23864" (UID: "2212e6a0-bdfa-41e1-8ff3-2b386ff23864"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.237664 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.237702 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.237720 4815 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.237734 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggqh7\" (UniqueName: \"kubernetes.io/projected/2212e6a0-bdfa-41e1-8ff3-2b386ff23864-kube-api-access-ggqh7\") on node \"crc\" DevicePath \"\"" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.699229 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" event={"ID":"2212e6a0-bdfa-41e1-8ff3-2b386ff23864","Type":"ContainerDied","Data":"44b8cb718468d0ab317fdd8c61f0868f55d5598b8d1656f9bb1c12166e7b1618"} Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.699508 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44b8cb718468d0ab317fdd8c61f0868f55d5598b8d1656f9bb1c12166e7b1618" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.699575 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.782378 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6"] Dec 05 09:33:41 crc kubenswrapper[4815]: E1205 09:33:41.782819 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2212e6a0-bdfa-41e1-8ff3-2b386ff23864" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.782843 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="2212e6a0-bdfa-41e1-8ff3-2b386ff23864" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.783047 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="2212e6a0-bdfa-41e1-8ff3-2b386ff23864" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.783712 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.785542 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.787142 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.787177 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.787866 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.800855 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6"] Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.848153 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spptt\" (UniqueName: \"kubernetes.io/projected/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-kube-api-access-spptt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.848433 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.848543 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.848814 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.951231 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spptt\" (UniqueName: \"kubernetes.io/projected/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-kube-api-access-spptt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.951295 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.951335 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.951403 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.957573 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.958167 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.959364 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:41 crc kubenswrapper[4815]: I1205 09:33:41.972554 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spptt\" (UniqueName: \"kubernetes.io/projected/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-kube-api-access-spptt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-925t6\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:42 crc kubenswrapper[4815]: I1205 09:33:42.099259 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:33:42 crc kubenswrapper[4815]: I1205 09:33:42.666226 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6"] Dec 05 09:33:42 crc kubenswrapper[4815]: W1205 09:33:42.670645 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c167339_94cc_4c3c_ae9c_0e9b0ceb256b.slice/crio-b9128749c1a147bec596a318b72c9923528757138ced29f6a398fc20c6b03f3f WatchSource:0}: Error finding container b9128749c1a147bec596a318b72c9923528757138ced29f6a398fc20c6b03f3f: Status 404 returned error can't find the container with id b9128749c1a147bec596a318b72c9923528757138ced29f6a398fc20c6b03f3f Dec 05 09:33:42 crc kubenswrapper[4815]: I1205 09:33:42.711096 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" event={"ID":"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b","Type":"ContainerStarted","Data":"b9128749c1a147bec596a318b72c9923528757138ced29f6a398fc20c6b03f3f"} Dec 05 09:33:43 crc kubenswrapper[4815]: I1205 09:33:43.722081 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" event={"ID":"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b","Type":"ContainerStarted","Data":"bc43e3d716f87a687679453b74e9ba1f7e6bd1ded31eb1d7e20d25467cc6015c"} Dec 05 09:33:43 crc kubenswrapper[4815]: I1205 09:33:43.746005 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" podStartSLOduration=2.186866161 podStartE2EDuration="2.745973149s" podCreationTimestamp="2025-12-05 09:33:41 +0000 UTC" firstStartedPulling="2025-12-05 09:33:42.673853297 +0000 UTC m=+1621.552460134" lastFinishedPulling="2025-12-05 09:33:43.232960285 +0000 UTC m=+1622.111567122" observedRunningTime="2025-12-05 09:33:43.740884818 +0000 UTC m=+1622.619491655" watchObservedRunningTime="2025-12-05 09:33:43.745973149 +0000 UTC m=+1622.624580006" Dec 05 09:33:50 crc kubenswrapper[4815]: E1205 09:33:50.846940 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2212e6a0_bdfa_41e1_8ff3_2b386ff23864.slice/crio-44b8cb718468d0ab317fdd8c61f0868f55d5598b8d1656f9bb1c12166e7b1618\": RecentStats: unable to find data in memory cache]" Dec 05 09:33:51 crc kubenswrapper[4815]: I1205 09:33:51.426204 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:33:51 crc kubenswrapper[4815]: E1205 09:33:51.426456 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:33:56 crc kubenswrapper[4815]: I1205 09:33:56.494195 4815 scope.go:117] "RemoveContainer" containerID="35a1af36a37de14e23c6772c3f8a10fe9ed1b8fdc6366f3afb7f69f07f8cd449" Dec 05 09:33:56 crc kubenswrapper[4815]: I1205 09:33:56.524176 4815 scope.go:117] "RemoveContainer" containerID="aaf49eeb4c0b5fa43b82eaae0c60cc6675ab0c58985d81f0bf994f967b6863d4" Dec 05 09:33:56 crc kubenswrapper[4815]: I1205 09:33:56.592909 4815 scope.go:117] "RemoveContainer" containerID="29e4959df45b3b0e9294e36efb913156e6274678794229bd25128e72e594fdad" Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.269440 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9xfdb"] Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.271674 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.294748 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9xfdb"] Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.379815 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-utilities\") pod \"community-operators-9xfdb\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.379868 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cx2t\" (UniqueName: \"kubernetes.io/projected/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-kube-api-access-9cx2t\") pod \"community-operators-9xfdb\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.380447 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-catalog-content\") pod \"community-operators-9xfdb\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.482466 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-catalog-content\") pod \"community-operators-9xfdb\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.482568 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-utilities\") pod \"community-operators-9xfdb\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.482601 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cx2t\" (UniqueName: \"kubernetes.io/projected/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-kube-api-access-9cx2t\") pod \"community-operators-9xfdb\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.483137 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-catalog-content\") pod \"community-operators-9xfdb\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.483164 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-utilities\") pod \"community-operators-9xfdb\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.508306 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cx2t\" (UniqueName: \"kubernetes.io/projected/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-kube-api-access-9cx2t\") pod \"community-operators-9xfdb\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:33:59 crc kubenswrapper[4815]: I1205 09:33:59.605768 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:34:00 crc kubenswrapper[4815]: I1205 09:34:00.161716 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9xfdb"] Dec 05 09:34:00 crc kubenswrapper[4815]: I1205 09:34:00.888718 4815 generic.go:334] "Generic (PLEG): container finished" podID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" containerID="4835b6329d00c7b29c3a510ea301a6974f871a32ff8d89d0df6e7a6a6481547c" exitCode=0 Dec 05 09:34:00 crc kubenswrapper[4815]: I1205 09:34:00.888774 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfdb" event={"ID":"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1","Type":"ContainerDied","Data":"4835b6329d00c7b29c3a510ea301a6974f871a32ff8d89d0df6e7a6a6481547c"} Dec 05 09:34:00 crc kubenswrapper[4815]: I1205 09:34:00.889036 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfdb" event={"ID":"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1","Type":"ContainerStarted","Data":"20c9f875bb2164f534ec0be7408026977a6884faf44ca358d25b8e4dd32e8be8"} Dec 05 09:34:00 crc kubenswrapper[4815]: I1205 09:34:00.890438 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:34:01 crc kubenswrapper[4815]: E1205 09:34:01.079714 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2212e6a0_bdfa_41e1_8ff3_2b386ff23864.slice/crio-44b8cb718468d0ab317fdd8c61f0868f55d5598b8d1656f9bb1c12166e7b1618\": RecentStats: unable to find data in memory cache]" Dec 05 09:34:01 crc kubenswrapper[4815]: I1205 09:34:01.901520 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfdb" event={"ID":"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1","Type":"ContainerStarted","Data":"af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef"} Dec 05 09:34:03 crc kubenswrapper[4815]: I1205 09:34:03.923229 4815 generic.go:334] "Generic (PLEG): container finished" podID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" containerID="af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef" exitCode=0 Dec 05 09:34:03 crc kubenswrapper[4815]: I1205 09:34:03.923389 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfdb" event={"ID":"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1","Type":"ContainerDied","Data":"af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef"} Dec 05 09:34:04 crc kubenswrapper[4815]: I1205 09:34:04.934445 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfdb" event={"ID":"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1","Type":"ContainerStarted","Data":"3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959"} Dec 05 09:34:04 crc kubenswrapper[4815]: I1205 09:34:04.960400 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9xfdb" podStartSLOduration=2.548291999 podStartE2EDuration="5.960377499s" podCreationTimestamp="2025-12-05 09:33:59 +0000 UTC" firstStartedPulling="2025-12-05 09:34:00.890144914 +0000 UTC m=+1639.768751751" lastFinishedPulling="2025-12-05 09:34:04.302230414 +0000 UTC m=+1643.180837251" observedRunningTime="2025-12-05 09:34:04.952082394 +0000 UTC m=+1643.830689231" watchObservedRunningTime="2025-12-05 09:34:04.960377499 +0000 UTC m=+1643.838984336" Dec 05 09:34:05 crc kubenswrapper[4815]: I1205 09:34:05.419681 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:34:05 crc kubenswrapper[4815]: E1205 09:34:05.420030 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:34:09 crc kubenswrapper[4815]: I1205 09:34:09.606135 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:34:09 crc kubenswrapper[4815]: I1205 09:34:09.606442 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:34:09 crc kubenswrapper[4815]: I1205 09:34:09.660446 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:34:10 crc kubenswrapper[4815]: I1205 09:34:10.023767 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:34:10 crc kubenswrapper[4815]: I1205 09:34:10.076182 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9xfdb"] Dec 05 09:34:11 crc kubenswrapper[4815]: E1205 09:34:11.315947 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2212e6a0_bdfa_41e1_8ff3_2b386ff23864.slice/crio-44b8cb718468d0ab317fdd8c61f0868f55d5598b8d1656f9bb1c12166e7b1618\": RecentStats: unable to find data in memory cache]" Dec 05 09:34:11 crc kubenswrapper[4815]: I1205 09:34:11.997216 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9xfdb" podUID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" containerName="registry-server" containerID="cri-o://3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959" gracePeriod=2 Dec 05 09:34:12 crc kubenswrapper[4815]: I1205 09:34:12.453518 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:34:12 crc kubenswrapper[4815]: I1205 09:34:12.526012 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9cx2t\" (UniqueName: \"kubernetes.io/projected/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-kube-api-access-9cx2t\") pod \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " Dec 05 09:34:12 crc kubenswrapper[4815]: I1205 09:34:12.526356 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-utilities\") pod \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " Dec 05 09:34:12 crc kubenswrapper[4815]: I1205 09:34:12.526527 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-catalog-content\") pod \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\" (UID: \"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1\") " Dec 05 09:34:12 crc kubenswrapper[4815]: I1205 09:34:12.527064 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-utilities" (OuterVolumeSpecName: "utilities") pod "13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" (UID: "13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:34:12 crc kubenswrapper[4815]: I1205 09:34:12.527457 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:34:12 crc kubenswrapper[4815]: I1205 09:34:12.534854 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-kube-api-access-9cx2t" (OuterVolumeSpecName: "kube-api-access-9cx2t") pod "13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" (UID: "13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1"). InnerVolumeSpecName "kube-api-access-9cx2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:34:12 crc kubenswrapper[4815]: I1205 09:34:12.584783 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" (UID: "13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:34:12 crc kubenswrapper[4815]: I1205 09:34:12.628924 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9cx2t\" (UniqueName: \"kubernetes.io/projected/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-kube-api-access-9cx2t\") on node \"crc\" DevicePath \"\"" Dec 05 09:34:12 crc kubenswrapper[4815]: I1205 09:34:12.628955 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.009680 4815 generic.go:334] "Generic (PLEG): container finished" podID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" containerID="3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959" exitCode=0 Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.009729 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfdb" event={"ID":"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1","Type":"ContainerDied","Data":"3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959"} Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.009762 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9xfdb" event={"ID":"13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1","Type":"ContainerDied","Data":"20c9f875bb2164f534ec0be7408026977a6884faf44ca358d25b8e4dd32e8be8"} Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.009781 4815 scope.go:117] "RemoveContainer" containerID="3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959" Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.009927 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9xfdb" Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.040285 4815 scope.go:117] "RemoveContainer" containerID="af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef" Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.069705 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9xfdb"] Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.080414 4815 scope.go:117] "RemoveContainer" containerID="4835b6329d00c7b29c3a510ea301a6974f871a32ff8d89d0df6e7a6a6481547c" Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.080612 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9xfdb"] Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.119331 4815 scope.go:117] "RemoveContainer" containerID="3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959" Dec 05 09:34:13 crc kubenswrapper[4815]: E1205 09:34:13.121884 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959\": container with ID starting with 3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959 not found: ID does not exist" containerID="3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959" Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.121949 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959"} err="failed to get container status \"3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959\": rpc error: code = NotFound desc = could not find container \"3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959\": container with ID starting with 3c419871d52cffc9a7bb8ce6eacc7ae7d686e7b02345c72e36fa44f2398ad959 not found: ID does not exist" Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.121997 4815 scope.go:117] "RemoveContainer" containerID="af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef" Dec 05 09:34:13 crc kubenswrapper[4815]: E1205 09:34:13.122411 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef\": container with ID starting with af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef not found: ID does not exist" containerID="af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef" Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.122460 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef"} err="failed to get container status \"af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef\": rpc error: code = NotFound desc = could not find container \"af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef\": container with ID starting with af143c49d45d10bde16cfa78f862d293dd6b52c00c10c9e0e4b61a5c88be4bef not found: ID does not exist" Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.122478 4815 scope.go:117] "RemoveContainer" containerID="4835b6329d00c7b29c3a510ea301a6974f871a32ff8d89d0df6e7a6a6481547c" Dec 05 09:34:13 crc kubenswrapper[4815]: E1205 09:34:13.122798 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4835b6329d00c7b29c3a510ea301a6974f871a32ff8d89d0df6e7a6a6481547c\": container with ID starting with 4835b6329d00c7b29c3a510ea301a6974f871a32ff8d89d0df6e7a6a6481547c not found: ID does not exist" containerID="4835b6329d00c7b29c3a510ea301a6974f871a32ff8d89d0df6e7a6a6481547c" Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.122826 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4835b6329d00c7b29c3a510ea301a6974f871a32ff8d89d0df6e7a6a6481547c"} err="failed to get container status \"4835b6329d00c7b29c3a510ea301a6974f871a32ff8d89d0df6e7a6a6481547c\": rpc error: code = NotFound desc = could not find container \"4835b6329d00c7b29c3a510ea301a6974f871a32ff8d89d0df6e7a6a6481547c\": container with ID starting with 4835b6329d00c7b29c3a510ea301a6974f871a32ff8d89d0df6e7a6a6481547c not found: ID does not exist" Dec 05 09:34:13 crc kubenswrapper[4815]: I1205 09:34:13.430626 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" path="/var/lib/kubelet/pods/13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1/volumes" Dec 05 09:34:18 crc kubenswrapper[4815]: I1205 09:34:18.418321 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:34:18 crc kubenswrapper[4815]: E1205 09:34:18.419083 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:34:21 crc kubenswrapper[4815]: E1205 09:34:21.579865 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2212e6a0_bdfa_41e1_8ff3_2b386ff23864.slice/crio-44b8cb718468d0ab317fdd8c61f0868f55d5598b8d1656f9bb1c12166e7b1618\": RecentStats: unable to find data in memory cache]" Dec 05 09:34:30 crc kubenswrapper[4815]: I1205 09:34:30.419102 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:34:30 crc kubenswrapper[4815]: E1205 09:34:30.420066 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:34:31 crc kubenswrapper[4815]: E1205 09:34:31.830226 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2212e6a0_bdfa_41e1_8ff3_2b386ff23864.slice/crio-44b8cb718468d0ab317fdd8c61f0868f55d5598b8d1656f9bb1c12166e7b1618\": RecentStats: unable to find data in memory cache]" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.649811 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tqnpk"] Dec 05 09:34:33 crc kubenswrapper[4815]: E1205 09:34:33.650452 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" containerName="extract-utilities" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.650474 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" containerName="extract-utilities" Dec 05 09:34:33 crc kubenswrapper[4815]: E1205 09:34:33.650524 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" containerName="registry-server" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.650532 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" containerName="registry-server" Dec 05 09:34:33 crc kubenswrapper[4815]: E1205 09:34:33.650553 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" containerName="extract-content" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.650559 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" containerName="extract-content" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.650774 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="13b77248-9f04-4ad1-8e0a-8e2b2a12c0f1" containerName="registry-server" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.652065 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.671077 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tqnpk"] Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.739653 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-catalog-content\") pod \"redhat-operators-tqnpk\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.739876 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-utilities\") pod \"redhat-operators-tqnpk\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.739922 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7kj7\" (UniqueName: \"kubernetes.io/projected/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-kube-api-access-h7kj7\") pod \"redhat-operators-tqnpk\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.841829 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7kj7\" (UniqueName: \"kubernetes.io/projected/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-kube-api-access-h7kj7\") pod \"redhat-operators-tqnpk\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.841901 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-catalog-content\") pod \"redhat-operators-tqnpk\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.842025 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-utilities\") pod \"redhat-operators-tqnpk\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.842440 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-utilities\") pod \"redhat-operators-tqnpk\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.843008 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-catalog-content\") pod \"redhat-operators-tqnpk\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.862296 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7kj7\" (UniqueName: \"kubernetes.io/projected/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-kube-api-access-h7kj7\") pod \"redhat-operators-tqnpk\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:33 crc kubenswrapper[4815]: I1205 09:34:33.988973 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:34 crc kubenswrapper[4815]: I1205 09:34:34.459689 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tqnpk"] Dec 05 09:34:35 crc kubenswrapper[4815]: I1205 09:34:35.254329 4815 generic.go:334] "Generic (PLEG): container finished" podID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerID="755d1ebc55b1204bc6b815efb44c5abf69711f32a9807f1f2e010244c33b8201" exitCode=0 Dec 05 09:34:35 crc kubenswrapper[4815]: I1205 09:34:35.254384 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnpk" event={"ID":"9c0311cc-cbcf-40f3-a31b-6c3c589afc79","Type":"ContainerDied","Data":"755d1ebc55b1204bc6b815efb44c5abf69711f32a9807f1f2e010244c33b8201"} Dec 05 09:34:35 crc kubenswrapper[4815]: I1205 09:34:35.254989 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnpk" event={"ID":"9c0311cc-cbcf-40f3-a31b-6c3c589afc79","Type":"ContainerStarted","Data":"970f3467168fbe1ee37529f3ec5ea55915c3678b8c1c88662c426a557f21a326"} Dec 05 09:34:37 crc kubenswrapper[4815]: I1205 09:34:37.274786 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnpk" event={"ID":"9c0311cc-cbcf-40f3-a31b-6c3c589afc79","Type":"ContainerStarted","Data":"d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01"} Dec 05 09:34:41 crc kubenswrapper[4815]: I1205 09:34:41.312197 4815 generic.go:334] "Generic (PLEG): container finished" podID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerID="d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01" exitCode=0 Dec 05 09:34:41 crc kubenswrapper[4815]: I1205 09:34:41.312285 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnpk" event={"ID":"9c0311cc-cbcf-40f3-a31b-6c3c589afc79","Type":"ContainerDied","Data":"d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01"} Dec 05 09:34:42 crc kubenswrapper[4815]: I1205 09:34:42.324096 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnpk" event={"ID":"9c0311cc-cbcf-40f3-a31b-6c3c589afc79","Type":"ContainerStarted","Data":"ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27"} Dec 05 09:34:42 crc kubenswrapper[4815]: I1205 09:34:42.350303 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tqnpk" podStartSLOduration=2.910933893 podStartE2EDuration="9.350253988s" podCreationTimestamp="2025-12-05 09:34:33 +0000 UTC" firstStartedPulling="2025-12-05 09:34:35.256244557 +0000 UTC m=+1674.134851394" lastFinishedPulling="2025-12-05 09:34:41.695564652 +0000 UTC m=+1680.574171489" observedRunningTime="2025-12-05 09:34:42.342097086 +0000 UTC m=+1681.220703923" watchObservedRunningTime="2025-12-05 09:34:42.350253988 +0000 UTC m=+1681.228860825" Dec 05 09:34:43 crc kubenswrapper[4815]: I1205 09:34:43.990941 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:43 crc kubenswrapper[4815]: I1205 09:34:43.992332 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:44 crc kubenswrapper[4815]: I1205 09:34:44.418579 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:34:44 crc kubenswrapper[4815]: E1205 09:34:44.418812 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:34:45 crc kubenswrapper[4815]: I1205 09:34:45.040953 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tqnpk" podUID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerName="registry-server" probeResult="failure" output=< Dec 05 09:34:45 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 09:34:45 crc kubenswrapper[4815]: > Dec 05 09:34:54 crc kubenswrapper[4815]: I1205 09:34:54.033231 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:54 crc kubenswrapper[4815]: I1205 09:34:54.083912 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:54 crc kubenswrapper[4815]: I1205 09:34:54.270666 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tqnpk"] Dec 05 09:34:55 crc kubenswrapper[4815]: I1205 09:34:55.454711 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tqnpk" podUID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerName="registry-server" containerID="cri-o://ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27" gracePeriod=2 Dec 05 09:34:55 crc kubenswrapper[4815]: I1205 09:34:55.909541 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.016081 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7kj7\" (UniqueName: \"kubernetes.io/projected/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-kube-api-access-h7kj7\") pod \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.016364 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-catalog-content\") pod \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.016471 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-utilities\") pod \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\" (UID: \"9c0311cc-cbcf-40f3-a31b-6c3c589afc79\") " Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.017630 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-utilities" (OuterVolumeSpecName: "utilities") pod "9c0311cc-cbcf-40f3-a31b-6c3c589afc79" (UID: "9c0311cc-cbcf-40f3-a31b-6c3c589afc79"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.018621 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.021566 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-kube-api-access-h7kj7" (OuterVolumeSpecName: "kube-api-access-h7kj7") pod "9c0311cc-cbcf-40f3-a31b-6c3c589afc79" (UID: "9c0311cc-cbcf-40f3-a31b-6c3c589afc79"). InnerVolumeSpecName "kube-api-access-h7kj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.120764 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7kj7\" (UniqueName: \"kubernetes.io/projected/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-kube-api-access-h7kj7\") on node \"crc\" DevicePath \"\"" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.127731 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9c0311cc-cbcf-40f3-a31b-6c3c589afc79" (UID: "9c0311cc-cbcf-40f3-a31b-6c3c589afc79"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.222131 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9c0311cc-cbcf-40f3-a31b-6c3c589afc79-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.467027 4815 generic.go:334] "Generic (PLEG): container finished" podID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerID="ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27" exitCode=0 Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.467081 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnpk" event={"ID":"9c0311cc-cbcf-40f3-a31b-6c3c589afc79","Type":"ContainerDied","Data":"ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27"} Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.467109 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tqnpk" event={"ID":"9c0311cc-cbcf-40f3-a31b-6c3c589afc79","Type":"ContainerDied","Data":"970f3467168fbe1ee37529f3ec5ea55915c3678b8c1c88662c426a557f21a326"} Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.467126 4815 scope.go:117] "RemoveContainer" containerID="ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.467245 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tqnpk" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.500971 4815 scope.go:117] "RemoveContainer" containerID="d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.504995 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tqnpk"] Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.517147 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tqnpk"] Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.533729 4815 scope.go:117] "RemoveContainer" containerID="755d1ebc55b1204bc6b815efb44c5abf69711f32a9807f1f2e010244c33b8201" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.568468 4815 scope.go:117] "RemoveContainer" containerID="ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27" Dec 05 09:34:56 crc kubenswrapper[4815]: E1205 09:34:56.569012 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27\": container with ID starting with ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27 not found: ID does not exist" containerID="ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.569054 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27"} err="failed to get container status \"ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27\": rpc error: code = NotFound desc = could not find container \"ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27\": container with ID starting with ebbfc1370c0a1c6e22a275f7ba26df21612064e81be15bfddf6c763036f63c27 not found: ID does not exist" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.569082 4815 scope.go:117] "RemoveContainer" containerID="d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01" Dec 05 09:34:56 crc kubenswrapper[4815]: E1205 09:34:56.569547 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01\": container with ID starting with d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01 not found: ID does not exist" containerID="d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.569571 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01"} err="failed to get container status \"d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01\": rpc error: code = NotFound desc = could not find container \"d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01\": container with ID starting with d37561be959bfe24c12a63e23da7e74ed77e3839d4fbdbc7d13b634c4f967e01 not found: ID does not exist" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.569588 4815 scope.go:117] "RemoveContainer" containerID="755d1ebc55b1204bc6b815efb44c5abf69711f32a9807f1f2e010244c33b8201" Dec 05 09:34:56 crc kubenswrapper[4815]: E1205 09:34:56.570047 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"755d1ebc55b1204bc6b815efb44c5abf69711f32a9807f1f2e010244c33b8201\": container with ID starting with 755d1ebc55b1204bc6b815efb44c5abf69711f32a9807f1f2e010244c33b8201 not found: ID does not exist" containerID="755d1ebc55b1204bc6b815efb44c5abf69711f32a9807f1f2e010244c33b8201" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.570071 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"755d1ebc55b1204bc6b815efb44c5abf69711f32a9807f1f2e010244c33b8201"} err="failed to get container status \"755d1ebc55b1204bc6b815efb44c5abf69711f32a9807f1f2e010244c33b8201\": rpc error: code = NotFound desc = could not find container \"755d1ebc55b1204bc6b815efb44c5abf69711f32a9807f1f2e010244c33b8201\": container with ID starting with 755d1ebc55b1204bc6b815efb44c5abf69711f32a9807f1f2e010244c33b8201 not found: ID does not exist" Dec 05 09:34:56 crc kubenswrapper[4815]: I1205 09:34:56.688041 4815 scope.go:117] "RemoveContainer" containerID="1f7ef1fbe6aa6478447e7bacb1027bef23a8b2e202b8cd1622fc1db108b8ce74" Dec 05 09:34:57 crc kubenswrapper[4815]: I1205 09:34:57.431041 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" path="/var/lib/kubelet/pods/9c0311cc-cbcf-40f3-a31b-6c3c589afc79/volumes" Dec 05 09:34:59 crc kubenswrapper[4815]: I1205 09:34:59.418982 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:34:59 crc kubenswrapper[4815]: E1205 09:34:59.419839 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:35:11 crc kubenswrapper[4815]: I1205 09:35:11.424070 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:35:11 crc kubenswrapper[4815]: E1205 09:35:11.424939 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:35:23 crc kubenswrapper[4815]: I1205 09:35:23.418254 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:35:23 crc kubenswrapper[4815]: E1205 09:35:23.418960 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:35:34 crc kubenswrapper[4815]: I1205 09:35:34.418716 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:35:34 crc kubenswrapper[4815]: E1205 09:35:34.419566 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:35:48 crc kubenswrapper[4815]: I1205 09:35:48.418629 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:35:48 crc kubenswrapper[4815]: E1205 09:35:48.419425 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:35:56 crc kubenswrapper[4815]: I1205 09:35:56.823587 4815 scope.go:117] "RemoveContainer" containerID="25adc14737ef33dd8e88043f554f20a7ecad51e478b30ff09764cfc3541be9a2" Dec 05 09:35:56 crc kubenswrapper[4815]: I1205 09:35:56.847000 4815 scope.go:117] "RemoveContainer" containerID="ee6b35a57697661062488bf6f8ea1bce91b1b01c51fa23980199e963b8b5eb7f" Dec 05 09:36:01 crc kubenswrapper[4815]: I1205 09:36:01.424362 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:36:01 crc kubenswrapper[4815]: E1205 09:36:01.426452 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:36:15 crc kubenswrapper[4815]: I1205 09:36:15.419595 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:36:15 crc kubenswrapper[4815]: E1205 09:36:15.420912 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:36:27 crc kubenswrapper[4815]: I1205 09:36:27.419012 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:36:27 crc kubenswrapper[4815]: E1205 09:36:27.419801 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:36:41 crc kubenswrapper[4815]: I1205 09:36:41.425311 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:36:41 crc kubenswrapper[4815]: E1205 09:36:41.427893 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:36:55 crc kubenswrapper[4815]: I1205 09:36:55.418647 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:36:55 crc kubenswrapper[4815]: E1205 09:36:55.420582 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:36:56 crc kubenswrapper[4815]: I1205 09:36:56.908852 4815 scope.go:117] "RemoveContainer" containerID="4c6fbd823baa707eca219357caadb96f1386f628ccfa02cd8cbe38810c2cc360" Dec 05 09:36:56 crc kubenswrapper[4815]: I1205 09:36:56.950883 4815 scope.go:117] "RemoveContainer" containerID="297c42c4df3d648eb7badc497ed93e52a142d1f33b5cae21b73a1cd8ed3de003" Dec 05 09:36:56 crc kubenswrapper[4815]: I1205 09:36:56.971923 4815 scope.go:117] "RemoveContainer" containerID="f3388f69e8652b0f560c12981de679df22d6cf1dbce24ccbb34dc0314322a52b" Dec 05 09:36:56 crc kubenswrapper[4815]: I1205 09:36:56.996850 4815 scope.go:117] "RemoveContainer" containerID="8011f7bff74e52e9fb6d190a39fa8c5709700fc614542e3dd4077d26ba85228c" Dec 05 09:37:06 crc kubenswrapper[4815]: I1205 09:37:06.419185 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:37:06 crc kubenswrapper[4815]: E1205 09:37:06.419938 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:37:17 crc kubenswrapper[4815]: I1205 09:37:17.418769 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:37:17 crc kubenswrapper[4815]: E1205 09:37:17.419507 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:37:26 crc kubenswrapper[4815]: I1205 09:37:26.049958 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-kf8dq"] Dec 05 09:37:26 crc kubenswrapper[4815]: I1205 09:37:26.065444 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-b445-account-create-update-84rcr"] Dec 05 09:37:26 crc kubenswrapper[4815]: I1205 09:37:26.086459 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-9wd7r"] Dec 05 09:37:26 crc kubenswrapper[4815]: I1205 09:37:26.094736 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-bb8e-account-create-update-fbh8q"] Dec 05 09:37:26 crc kubenswrapper[4815]: I1205 09:37:26.102902 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-kf8dq"] Dec 05 09:37:26 crc kubenswrapper[4815]: I1205 09:37:26.111105 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-b445-account-create-update-84rcr"] Dec 05 09:37:26 crc kubenswrapper[4815]: I1205 09:37:26.119867 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-bb8e-account-create-update-fbh8q"] Dec 05 09:37:26 crc kubenswrapper[4815]: I1205 09:37:26.128093 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-9wd7r"] Dec 05 09:37:27 crc kubenswrapper[4815]: I1205 09:37:27.430311 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd" path="/var/lib/kubelet/pods/7bb0bc8f-eecf-4488-ab4e-199ce7afd5dd/volumes" Dec 05 09:37:27 crc kubenswrapper[4815]: I1205 09:37:27.431749 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bdc590e0-1c3a-418e-828b-eb43990319de" path="/var/lib/kubelet/pods/bdc590e0-1c3a-418e-828b-eb43990319de/volumes" Dec 05 09:37:27 crc kubenswrapper[4815]: I1205 09:37:27.432547 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf0e72c3-db89-4255-9c0b-50734039276b" path="/var/lib/kubelet/pods/cf0e72c3-db89-4255-9c0b-50734039276b/volumes" Dec 05 09:37:27 crc kubenswrapper[4815]: I1205 09:37:27.433300 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e31d5e84-0c31-4e29-a234-3d2ba4c45c3e" path="/var/lib/kubelet/pods/e31d5e84-0c31-4e29-a234-3d2ba4c45c3e/volumes" Dec 05 09:37:32 crc kubenswrapper[4815]: I1205 09:37:32.077189 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-bjzq2"] Dec 05 09:37:32 crc kubenswrapper[4815]: I1205 09:37:32.086706 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-3451-account-create-update-thcm9"] Dec 05 09:37:32 crc kubenswrapper[4815]: I1205 09:37:32.094090 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-bjzq2"] Dec 05 09:37:32 crc kubenswrapper[4815]: I1205 09:37:32.102964 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-3451-account-create-update-thcm9"] Dec 05 09:37:32 crc kubenswrapper[4815]: I1205 09:37:32.418331 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:37:32 crc kubenswrapper[4815]: E1205 09:37:32.418594 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:37:33 crc kubenswrapper[4815]: I1205 09:37:33.432975 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3231e31e-80a7-4aec-a346-2d34f455b62a" path="/var/lib/kubelet/pods/3231e31e-80a7-4aec-a346-2d34f455b62a/volumes" Dec 05 09:37:33 crc kubenswrapper[4815]: I1205 09:37:33.434218 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9170f00-6a0c-45db-86ca-906ed973f9bb" path="/var/lib/kubelet/pods/f9170f00-6a0c-45db-86ca-906ed973f9bb/volumes" Dec 05 09:37:45 crc kubenswrapper[4815]: I1205 09:37:45.418769 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:37:45 crc kubenswrapper[4815]: E1205 09:37:45.419405 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:37:46 crc kubenswrapper[4815]: I1205 09:37:46.184170 4815 generic.go:334] "Generic (PLEG): container finished" podID="7c167339-94cc-4c3c-ae9c-0e9b0ceb256b" containerID="bc43e3d716f87a687679453b74e9ba1f7e6bd1ded31eb1d7e20d25467cc6015c" exitCode=0 Dec 05 09:37:46 crc kubenswrapper[4815]: I1205 09:37:46.184216 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" event={"ID":"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b","Type":"ContainerDied","Data":"bc43e3d716f87a687679453b74e9ba1f7e6bd1ded31eb1d7e20d25467cc6015c"} Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.554089 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.655365 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-bootstrap-combined-ca-bundle\") pod \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.655716 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spptt\" (UniqueName: \"kubernetes.io/projected/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-kube-api-access-spptt\") pod \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.655826 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-inventory\") pod \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.655966 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-ssh-key\") pod \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\" (UID: \"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b\") " Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.661249 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "7c167339-94cc-4c3c-ae9c-0e9b0ceb256b" (UID: "7c167339-94cc-4c3c-ae9c-0e9b0ceb256b"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.668769 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-kube-api-access-spptt" (OuterVolumeSpecName: "kube-api-access-spptt") pod "7c167339-94cc-4c3c-ae9c-0e9b0ceb256b" (UID: "7c167339-94cc-4c3c-ae9c-0e9b0ceb256b"). InnerVolumeSpecName "kube-api-access-spptt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.689620 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-inventory" (OuterVolumeSpecName: "inventory") pod "7c167339-94cc-4c3c-ae9c-0e9b0ceb256b" (UID: "7c167339-94cc-4c3c-ae9c-0e9b0ceb256b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.689671 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7c167339-94cc-4c3c-ae9c-0e9b0ceb256b" (UID: "7c167339-94cc-4c3c-ae9c-0e9b0ceb256b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.757979 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.758019 4815 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.758032 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spptt\" (UniqueName: \"kubernetes.io/projected/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-kube-api-access-spptt\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:47 crc kubenswrapper[4815]: I1205 09:37:47.758041 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.203269 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" event={"ID":"7c167339-94cc-4c3c-ae9c-0e9b0ceb256b","Type":"ContainerDied","Data":"b9128749c1a147bec596a318b72c9923528757138ced29f6a398fc20c6b03f3f"} Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.203316 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9128749c1a147bec596a318b72c9923528757138ced29f6a398fc20c6b03f3f" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.203299 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.297502 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj"] Dec 05 09:37:48 crc kubenswrapper[4815]: E1205 09:37:48.298027 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerName="extract-utilities" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.298080 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerName="extract-utilities" Dec 05 09:37:48 crc kubenswrapper[4815]: E1205 09:37:48.298108 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerName="extract-content" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.298117 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerName="extract-content" Dec 05 09:37:48 crc kubenswrapper[4815]: E1205 09:37:48.298131 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c167339-94cc-4c3c-ae9c-0e9b0ceb256b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.298141 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c167339-94cc-4c3c-ae9c-0e9b0ceb256b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 05 09:37:48 crc kubenswrapper[4815]: E1205 09:37:48.298159 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerName="registry-server" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.298167 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerName="registry-server" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.298405 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c0311cc-cbcf-40f3-a31b-6c3c589afc79" containerName="registry-server" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.298426 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c167339-94cc-4c3c-ae9c-0e9b0ceb256b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.299176 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.301737 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.301955 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.302193 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.316419 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj"] Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.316614 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.471539 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hslnj\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.471859 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hslnj\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.472053 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmr2p\" (UniqueName: \"kubernetes.io/projected/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-kube-api-access-qmr2p\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hslnj\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.574080 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hslnj\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.574178 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hslnj\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.574227 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmr2p\" (UniqueName: \"kubernetes.io/projected/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-kube-api-access-qmr2p\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hslnj\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.578383 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hslnj\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.578409 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hslnj\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.594876 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmr2p\" (UniqueName: \"kubernetes.io/projected/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-kube-api-access-qmr2p\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-hslnj\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:37:48 crc kubenswrapper[4815]: I1205 09:37:48.619066 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:37:49 crc kubenswrapper[4815]: I1205 09:37:49.123444 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj"] Dec 05 09:37:49 crc kubenswrapper[4815]: I1205 09:37:49.212658 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" event={"ID":"c33d38f4-80e3-421d-ad3c-19fd6d7005f8","Type":"ContainerStarted","Data":"507addd9d9c527b4bb89c087cc82d2448d435cd36e0acc30b052a4fcadd13c53"} Dec 05 09:37:50 crc kubenswrapper[4815]: I1205 09:37:50.237034 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" event={"ID":"c33d38f4-80e3-421d-ad3c-19fd6d7005f8","Type":"ContainerStarted","Data":"2566e67858c16d008dcca1b1cac59338167110b391114555b4637ead1be6f595"} Dec 05 09:37:50 crc kubenswrapper[4815]: I1205 09:37:50.282401 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" podStartSLOduration=1.687868208 podStartE2EDuration="2.282365763s" podCreationTimestamp="2025-12-05 09:37:48 +0000 UTC" firstStartedPulling="2025-12-05 09:37:49.131098621 +0000 UTC m=+1868.009705458" lastFinishedPulling="2025-12-05 09:37:49.725596176 +0000 UTC m=+1868.604203013" observedRunningTime="2025-12-05 09:37:50.263878534 +0000 UTC m=+1869.142485401" watchObservedRunningTime="2025-12-05 09:37:50.282365763 +0000 UTC m=+1869.160972600" Dec 05 09:37:57 crc kubenswrapper[4815]: I1205 09:37:57.055458 4815 scope.go:117] "RemoveContainer" containerID="e3cb45ebf1c6ef6bfe6cf41744782bd99fea5008dd21d0fe01295c1ea2d23829" Dec 05 09:37:57 crc kubenswrapper[4815]: I1205 09:37:57.079080 4815 scope.go:117] "RemoveContainer" containerID="24e69df98780efdeaa978b1f4f10219cbb3516162678605a6c91f1fb00c6dd99" Dec 05 09:37:57 crc kubenswrapper[4815]: I1205 09:37:57.128422 4815 scope.go:117] "RemoveContainer" containerID="00f59d79f53768ab866bbd7fc7f3d51a947442e5063b5900d31021c4f1baedc8" Dec 05 09:37:57 crc kubenswrapper[4815]: I1205 09:37:57.277828 4815 scope.go:117] "RemoveContainer" containerID="3cef96d92575e8674e0505193a7587c99e8c0447f2e5da910606ca6413cc741f" Dec 05 09:37:57 crc kubenswrapper[4815]: I1205 09:37:57.306816 4815 scope.go:117] "RemoveContainer" containerID="bae28e56dcceb530d59557ad2918bdf0be3088cb81734c3dde78d8a83654a904" Dec 05 09:37:57 crc kubenswrapper[4815]: I1205 09:37:57.396844 4815 scope.go:117] "RemoveContainer" containerID="73609fb74300fe5e6c337116b71b5423ea6d50fd734c03b5193239a0834a646e" Dec 05 09:37:57 crc kubenswrapper[4815]: I1205 09:37:57.435915 4815 scope.go:117] "RemoveContainer" containerID="6752875772ae7fecf27a3d608a95ec2fa45a5d6648ebcf06b347addc2ed51eb2" Dec 05 09:37:57 crc kubenswrapper[4815]: I1205 09:37:57.481452 4815 scope.go:117] "RemoveContainer" containerID="dbee94149eacb41f50efcef2b2fd7b41b04dc3f2a2b1bd52e79baccbc4bf0b1e" Dec 05 09:37:59 crc kubenswrapper[4815]: I1205 09:37:59.418927 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:37:59 crc kubenswrapper[4815]: E1205 09:37:59.419515 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.049741 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-6pbnl"] Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.057721 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-z6t49"] Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.069436 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-6pbnl"] Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.078631 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-2qpg9"] Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.088540 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-z6t49"] Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.095523 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-63a3-account-create-update-5hprm"] Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.102615 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-2qpg9"] Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.110704 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-63a3-account-create-update-5hprm"] Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.427849 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a6fff39-63be-4b79-a9a0-dbcf8162a231" path="/var/lib/kubelet/pods/2a6fff39-63be-4b79-a9a0-dbcf8162a231/volumes" Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.428835 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45462c49-fb2a-48e8-bd00-98efa3a73da7" path="/var/lib/kubelet/pods/45462c49-fb2a-48e8-bd00-98efa3a73da7/volumes" Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.429522 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dfe6e19-f7c0-4360-945c-e70be5e60932" path="/var/lib/kubelet/pods/6dfe6e19-f7c0-4360-945c-e70be5e60932/volumes" Dec 05 09:38:01 crc kubenswrapper[4815]: I1205 09:38:01.430132 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7a77236-882a-4a90-99bc-b74679d44e01" path="/var/lib/kubelet/pods/b7a77236-882a-4a90-99bc-b74679d44e01/volumes" Dec 05 09:38:02 crc kubenswrapper[4815]: I1205 09:38:02.046191 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-c640-account-create-update-pb7l7"] Dec 05 09:38:02 crc kubenswrapper[4815]: I1205 09:38:02.053862 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-d057-account-create-update-kdmmk"] Dec 05 09:38:02 crc kubenswrapper[4815]: I1205 09:38:02.061394 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-c640-account-create-update-pb7l7"] Dec 05 09:38:02 crc kubenswrapper[4815]: I1205 09:38:02.068660 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-d057-account-create-update-kdmmk"] Dec 05 09:38:03 crc kubenswrapper[4815]: I1205 09:38:03.429258 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d044ebe-1e9f-4412-a94f-702501190349" path="/var/lib/kubelet/pods/9d044ebe-1e9f-4412-a94f-702501190349/volumes" Dec 05 09:38:03 crc kubenswrapper[4815]: I1205 09:38:03.430302 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8373b63-0203-47e7-b23c-714afe735109" path="/var/lib/kubelet/pods/f8373b63-0203-47e7-b23c-714afe735109/volumes" Dec 05 09:38:10 crc kubenswrapper[4815]: I1205 09:38:10.419677 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:38:10 crc kubenswrapper[4815]: E1205 09:38:10.420462 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:38:19 crc kubenswrapper[4815]: I1205 09:38:19.029235 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-bmkjh"] Dec 05 09:38:19 crc kubenswrapper[4815]: I1205 09:38:19.036715 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-bmkjh"] Dec 05 09:38:19 crc kubenswrapper[4815]: I1205 09:38:19.432351 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a86ad8ca-5199-41c4-ab4e-22826ea4b65d" path="/var/lib/kubelet/pods/a86ad8ca-5199-41c4-ab4e-22826ea4b65d/volumes" Dec 05 09:38:23 crc kubenswrapper[4815]: I1205 09:38:23.418792 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:38:23 crc kubenswrapper[4815]: I1205 09:38:23.664838 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"1db619e28f79e22d68fc8699b12d64acf5e1c448e650971b8846c7b1098b9c14"} Dec 05 09:38:50 crc kubenswrapper[4815]: I1205 09:38:50.045170 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-wjc6t"] Dec 05 09:38:50 crc kubenswrapper[4815]: I1205 09:38:50.053649 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-wjc6t"] Dec 05 09:38:51 crc kubenswrapper[4815]: I1205 09:38:51.450548 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d00433e1-638a-4b77-bd8b-9052d9c9bc11" path="/var/lib/kubelet/pods/d00433e1-638a-4b77-bd8b-9052d9c9bc11/volumes" Dec 05 09:38:57 crc kubenswrapper[4815]: I1205 09:38:57.636873 4815 scope.go:117] "RemoveContainer" containerID="22c4d51a44d918028a38229aaac9675684db6f88b5b01bfc62ce0fc9b2becfed" Dec 05 09:38:57 crc kubenswrapper[4815]: I1205 09:38:57.709316 4815 scope.go:117] "RemoveContainer" containerID="db0e92c9f5a41f6029013e1bdb3c8547ec232497e34d12d03d0a488144286229" Dec 05 09:38:57 crc kubenswrapper[4815]: I1205 09:38:57.778346 4815 scope.go:117] "RemoveContainer" containerID="1b7e1ff9d5277936ebb7ec3ecfdeca681e7b2c8578088537eca5c2cde35f048f" Dec 05 09:38:57 crc kubenswrapper[4815]: I1205 09:38:57.824563 4815 scope.go:117] "RemoveContainer" containerID="41337d3bda688a2910f9311369522d3ffcf27c4c3d69ed3bce373f33009afaba" Dec 05 09:38:57 crc kubenswrapper[4815]: I1205 09:38:57.890160 4815 scope.go:117] "RemoveContainer" containerID="7dfe05922f1b2c3eac292e7c7d589817b18eba9a50d15a97ad9db4e6beb2f1b2" Dec 05 09:38:57 crc kubenswrapper[4815]: I1205 09:38:57.926723 4815 scope.go:117] "RemoveContainer" containerID="cf891be6ee412a5b3b684ebcf7a903fdcbf3673fb15d96e38af7b05544c02c7b" Dec 05 09:38:57 crc kubenswrapper[4815]: I1205 09:38:57.971196 4815 scope.go:117] "RemoveContainer" containerID="c984a6a29a187f2bd2e3700a3572db64cb3aa0b921cf09c427888a1e39d63a8b" Dec 05 09:38:58 crc kubenswrapper[4815]: I1205 09:38:58.021238 4815 scope.go:117] "RemoveContainer" containerID="dda28bff0f5484c532faafe0684bae0d1d724980f07b9c53bff78ad1fe077a11" Dec 05 09:38:59 crc kubenswrapper[4815]: I1205 09:38:59.034177 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-4sfv2"] Dec 05 09:38:59 crc kubenswrapper[4815]: I1205 09:38:59.043225 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-4sfv2"] Dec 05 09:38:59 crc kubenswrapper[4815]: I1205 09:38:59.056979 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-59q99"] Dec 05 09:38:59 crc kubenswrapper[4815]: I1205 09:38:59.065533 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-59q99"] Dec 05 09:38:59 crc kubenswrapper[4815]: I1205 09:38:59.431112 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b7b7063-a8d5-41bd-b969-72035e0f2c6b" path="/var/lib/kubelet/pods/1b7b7063-a8d5-41bd-b969-72035e0f2c6b/volumes" Dec 05 09:38:59 crc kubenswrapper[4815]: I1205 09:38:59.432245 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="664c6884-a071-4e07-a90a-14146b1bb46e" path="/var/lib/kubelet/pods/664c6884-a071-4e07-a90a-14146b1bb46e/volumes" Dec 05 09:39:16 crc kubenswrapper[4815]: I1205 09:39:16.250301 4815 generic.go:334] "Generic (PLEG): container finished" podID="c33d38f4-80e3-421d-ad3c-19fd6d7005f8" containerID="2566e67858c16d008dcca1b1cac59338167110b391114555b4637ead1be6f595" exitCode=0 Dec 05 09:39:16 crc kubenswrapper[4815]: I1205 09:39:16.250365 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" event={"ID":"c33d38f4-80e3-421d-ad3c-19fd6d7005f8","Type":"ContainerDied","Data":"2566e67858c16d008dcca1b1cac59338167110b391114555b4637ead1be6f595"} Dec 05 09:39:17 crc kubenswrapper[4815]: I1205 09:39:17.659157 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:39:17 crc kubenswrapper[4815]: I1205 09:39:17.807394 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-inventory\") pod \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " Dec 05 09:39:17 crc kubenswrapper[4815]: I1205 09:39:17.807879 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-ssh-key\") pod \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " Dec 05 09:39:17 crc kubenswrapper[4815]: I1205 09:39:17.808114 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmr2p\" (UniqueName: \"kubernetes.io/projected/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-kube-api-access-qmr2p\") pod \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\" (UID: \"c33d38f4-80e3-421d-ad3c-19fd6d7005f8\") " Dec 05 09:39:17 crc kubenswrapper[4815]: I1205 09:39:17.820883 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-kube-api-access-qmr2p" (OuterVolumeSpecName: "kube-api-access-qmr2p") pod "c33d38f4-80e3-421d-ad3c-19fd6d7005f8" (UID: "c33d38f4-80e3-421d-ad3c-19fd6d7005f8"). InnerVolumeSpecName "kube-api-access-qmr2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:39:17 crc kubenswrapper[4815]: I1205 09:39:17.838779 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c33d38f4-80e3-421d-ad3c-19fd6d7005f8" (UID: "c33d38f4-80e3-421d-ad3c-19fd6d7005f8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:39:17 crc kubenswrapper[4815]: I1205 09:39:17.838863 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-inventory" (OuterVolumeSpecName: "inventory") pod "c33d38f4-80e3-421d-ad3c-19fd6d7005f8" (UID: "c33d38f4-80e3-421d-ad3c-19fd6d7005f8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:39:17 crc kubenswrapper[4815]: I1205 09:39:17.911947 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:39:17 crc kubenswrapper[4815]: I1205 09:39:17.912000 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:39:17 crc kubenswrapper[4815]: I1205 09:39:17.912019 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmr2p\" (UniqueName: \"kubernetes.io/projected/c33d38f4-80e3-421d-ad3c-19fd6d7005f8-kube-api-access-qmr2p\") on node \"crc\" DevicePath \"\"" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.268209 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" event={"ID":"c33d38f4-80e3-421d-ad3c-19fd6d7005f8","Type":"ContainerDied","Data":"507addd9d9c527b4bb89c087cc82d2448d435cd36e0acc30b052a4fcadd13c53"} Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.268252 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="507addd9d9c527b4bb89c087cc82d2448d435cd36e0acc30b052a4fcadd13c53" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.268252 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.361499 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p"] Dec 05 09:39:18 crc kubenswrapper[4815]: E1205 09:39:18.361984 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c33d38f4-80e3-421d-ad3c-19fd6d7005f8" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.362018 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c33d38f4-80e3-421d-ad3c-19fd6d7005f8" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.362221 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c33d38f4-80e3-421d-ad3c-19fd6d7005f8" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.363038 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.366915 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.366928 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.367536 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.368206 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.376790 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p"] Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.526126 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.526646 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.527020 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbp8g\" (UniqueName: \"kubernetes.io/projected/8b97a436-14ba-4b10-91b9-d7d90b73c62d-kube-api-access-nbp8g\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.628579 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbp8g\" (UniqueName: \"kubernetes.io/projected/8b97a436-14ba-4b10-91b9-d7d90b73c62d-kube-api-access-nbp8g\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.628698 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.628764 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.644403 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.646961 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.652027 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbp8g\" (UniqueName: \"kubernetes.io/projected/8b97a436-14ba-4b10-91b9-d7d90b73c62d-kube-api-access-nbp8g\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:18 crc kubenswrapper[4815]: I1205 09:39:18.680170 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:19 crc kubenswrapper[4815]: I1205 09:39:19.233221 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p"] Dec 05 09:39:19 crc kubenswrapper[4815]: I1205 09:39:19.238751 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:39:19 crc kubenswrapper[4815]: I1205 09:39:19.276531 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" event={"ID":"8b97a436-14ba-4b10-91b9-d7d90b73c62d","Type":"ContainerStarted","Data":"c6528beb913a8bae4adc70cd66c204edac46854e1d23dd21796f4e3f81ed4e09"} Dec 05 09:39:20 crc kubenswrapper[4815]: I1205 09:39:20.290416 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" event={"ID":"8b97a436-14ba-4b10-91b9-d7d90b73c62d","Type":"ContainerStarted","Data":"05862728ea7ea39070e64b9e68cd5c3ade5b4016914f2ad349ccb13c78a598bd"} Dec 05 09:39:20 crc kubenswrapper[4815]: I1205 09:39:20.326956 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" podStartSLOduration=1.9010797240000001 podStartE2EDuration="2.326927404s" podCreationTimestamp="2025-12-05 09:39:18 +0000 UTC" firstStartedPulling="2025-12-05 09:39:19.238474582 +0000 UTC m=+1958.117081419" lastFinishedPulling="2025-12-05 09:39:19.664322262 +0000 UTC m=+1958.542929099" observedRunningTime="2025-12-05 09:39:20.306460089 +0000 UTC m=+1959.185066936" watchObservedRunningTime="2025-12-05 09:39:20.326927404 +0000 UTC m=+1959.205534261" Dec 05 09:39:23 crc kubenswrapper[4815]: I1205 09:39:23.045264 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-dgjpz"] Dec 05 09:39:23 crc kubenswrapper[4815]: I1205 09:39:23.056870 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-ss9wk"] Dec 05 09:39:23 crc kubenswrapper[4815]: I1205 09:39:23.064652 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-dgjpz"] Dec 05 09:39:23 crc kubenswrapper[4815]: I1205 09:39:23.073117 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-ss9wk"] Dec 05 09:39:23 crc kubenswrapper[4815]: I1205 09:39:23.438658 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2ea4b20-281a-43e1-b6af-c2c209b10e9d" path="/var/lib/kubelet/pods/a2ea4b20-281a-43e1-b6af-c2c209b10e9d/volumes" Dec 05 09:39:23 crc kubenswrapper[4815]: I1205 09:39:23.440001 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c66190b5-87b0-440b-9894-1bffb2cfca3f" path="/var/lib/kubelet/pods/c66190b5-87b0-440b-9894-1bffb2cfca3f/volumes" Dec 05 09:39:24 crc kubenswrapper[4815]: I1205 09:39:24.028867 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-96bs8"] Dec 05 09:39:24 crc kubenswrapper[4815]: I1205 09:39:24.035564 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-96bs8"] Dec 05 09:39:25 crc kubenswrapper[4815]: I1205 09:39:25.337251 4815 generic.go:334] "Generic (PLEG): container finished" podID="8b97a436-14ba-4b10-91b9-d7d90b73c62d" containerID="05862728ea7ea39070e64b9e68cd5c3ade5b4016914f2ad349ccb13c78a598bd" exitCode=0 Dec 05 09:39:25 crc kubenswrapper[4815]: I1205 09:39:25.337327 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" event={"ID":"8b97a436-14ba-4b10-91b9-d7d90b73c62d","Type":"ContainerDied","Data":"05862728ea7ea39070e64b9e68cd5c3ade5b4016914f2ad349ccb13c78a598bd"} Dec 05 09:39:25 crc kubenswrapper[4815]: I1205 09:39:25.431462 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c832b179-b3a7-45d6-a360-11e05100d1f2" path="/var/lib/kubelet/pods/c832b179-b3a7-45d6-a360-11e05100d1f2/volumes" Dec 05 09:39:26 crc kubenswrapper[4815]: I1205 09:39:26.842595 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:26 crc kubenswrapper[4815]: I1205 09:39:26.920031 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-ssh-key\") pod \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " Dec 05 09:39:26 crc kubenswrapper[4815]: I1205 09:39:26.920616 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-inventory\") pod \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " Dec 05 09:39:26 crc kubenswrapper[4815]: I1205 09:39:26.920764 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbp8g\" (UniqueName: \"kubernetes.io/projected/8b97a436-14ba-4b10-91b9-d7d90b73c62d-kube-api-access-nbp8g\") pod \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\" (UID: \"8b97a436-14ba-4b10-91b9-d7d90b73c62d\") " Dec 05 09:39:26 crc kubenswrapper[4815]: I1205 09:39:26.940557 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b97a436-14ba-4b10-91b9-d7d90b73c62d-kube-api-access-nbp8g" (OuterVolumeSpecName: "kube-api-access-nbp8g") pod "8b97a436-14ba-4b10-91b9-d7d90b73c62d" (UID: "8b97a436-14ba-4b10-91b9-d7d90b73c62d"). InnerVolumeSpecName "kube-api-access-nbp8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:39:26 crc kubenswrapper[4815]: I1205 09:39:26.970505 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-inventory" (OuterVolumeSpecName: "inventory") pod "8b97a436-14ba-4b10-91b9-d7d90b73c62d" (UID: "8b97a436-14ba-4b10-91b9-d7d90b73c62d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:39:26 crc kubenswrapper[4815]: I1205 09:39:26.971051 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8b97a436-14ba-4b10-91b9-d7d90b73c62d" (UID: "8b97a436-14ba-4b10-91b9-d7d90b73c62d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.022037 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.022065 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbp8g\" (UniqueName: \"kubernetes.io/projected/8b97a436-14ba-4b10-91b9-d7d90b73c62d-kube-api-access-nbp8g\") on node \"crc\" DevicePath \"\"" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.022077 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8b97a436-14ba-4b10-91b9-d7d90b73c62d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.356013 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" event={"ID":"8b97a436-14ba-4b10-91b9-d7d90b73c62d","Type":"ContainerDied","Data":"c6528beb913a8bae4adc70cd66c204edac46854e1d23dd21796f4e3f81ed4e09"} Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.356059 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c6528beb913a8bae4adc70cd66c204edac46854e1d23dd21796f4e3f81ed4e09" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.356087 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.459042 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9"] Dec 05 09:39:27 crc kubenswrapper[4815]: E1205 09:39:27.459688 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b97a436-14ba-4b10-91b9-d7d90b73c62d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.459705 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b97a436-14ba-4b10-91b9-d7d90b73c62d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.459911 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b97a436-14ba-4b10-91b9-d7d90b73c62d" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.460527 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.467912 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.468218 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.468608 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.469576 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.487231 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9"] Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.631524 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qcxk9\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.631610 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wwsd\" (UniqueName: \"kubernetes.io/projected/bf14d613-f583-4916-9a10-07fe2de25e48-kube-api-access-2wwsd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qcxk9\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.631674 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qcxk9\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.734297 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qcxk9\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.734375 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wwsd\" (UniqueName: \"kubernetes.io/projected/bf14d613-f583-4916-9a10-07fe2de25e48-kube-api-access-2wwsd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qcxk9\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.734418 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qcxk9\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.738375 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qcxk9\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.744081 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qcxk9\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.757645 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wwsd\" (UniqueName: \"kubernetes.io/projected/bf14d613-f583-4916-9a10-07fe2de25e48-kube-api-access-2wwsd\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qcxk9\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:39:27 crc kubenswrapper[4815]: I1205 09:39:27.788438 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:39:28 crc kubenswrapper[4815]: I1205 09:39:28.307655 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9"] Dec 05 09:39:28 crc kubenswrapper[4815]: I1205 09:39:28.363576 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" event={"ID":"bf14d613-f583-4916-9a10-07fe2de25e48","Type":"ContainerStarted","Data":"0ddff0c6b7161d7a2a1bf5a1a2c9d17de9da3e24c4b9d7e8756cdb3c7b977e54"} Dec 05 09:39:29 crc kubenswrapper[4815]: I1205 09:39:29.372778 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" event={"ID":"bf14d613-f583-4916-9a10-07fe2de25e48","Type":"ContainerStarted","Data":"ff883880d2cd98b119cf093dcfe70e10b93acbe38cca5a0f6b807a841aa2903b"} Dec 05 09:39:29 crc kubenswrapper[4815]: I1205 09:39:29.393347 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" podStartSLOduration=1.8582606670000001 podStartE2EDuration="2.39332509s" podCreationTimestamp="2025-12-05 09:39:27 +0000 UTC" firstStartedPulling="2025-12-05 09:39:28.308933807 +0000 UTC m=+1967.187540644" lastFinishedPulling="2025-12-05 09:39:28.84399823 +0000 UTC m=+1967.722605067" observedRunningTime="2025-12-05 09:39:29.389213108 +0000 UTC m=+1968.267819955" watchObservedRunningTime="2025-12-05 09:39:29.39332509 +0000 UTC m=+1968.271931937" Dec 05 09:39:58 crc kubenswrapper[4815]: I1205 09:39:58.226459 4815 scope.go:117] "RemoveContainer" containerID="ac13c63149b7164d533764020d964e0736c593645c983b6e58b63ac587517a86" Dec 05 09:39:58 crc kubenswrapper[4815]: I1205 09:39:58.258403 4815 scope.go:117] "RemoveContainer" containerID="c86c9131a647365e140b7d211616e5b78f71e312cf85293d078ba9093f0e52da" Dec 05 09:39:58 crc kubenswrapper[4815]: I1205 09:39:58.329529 4815 scope.go:117] "RemoveContainer" containerID="d460b7b15ce45f9ce88ec5a4df9c5bfd875de498a3ecb6315d71b8d6fc916d84" Dec 05 09:39:58 crc kubenswrapper[4815]: I1205 09:39:58.379430 4815 scope.go:117] "RemoveContainer" containerID="0bd36f2c0a25614d8e3b243d5764d3f2b92f0ea7796d9d90abf1961264d67e81" Dec 05 09:39:58 crc kubenswrapper[4815]: I1205 09:39:58.408953 4815 scope.go:117] "RemoveContainer" containerID="d09086a87a983214a6a3881c0aa175532e712f484214c0057dc7f8fea016f095" Dec 05 09:40:03 crc kubenswrapper[4815]: I1205 09:40:03.051239 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-2bb0-account-create-update-f4zbp"] Dec 05 09:40:03 crc kubenswrapper[4815]: I1205 09:40:03.059122 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-jmvvv"] Dec 05 09:40:03 crc kubenswrapper[4815]: I1205 09:40:03.068078 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-2bb0-account-create-update-f4zbp"] Dec 05 09:40:03 crc kubenswrapper[4815]: I1205 09:40:03.075691 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-jmvvv"] Dec 05 09:40:03 crc kubenswrapper[4815]: I1205 09:40:03.437664 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d48f238-7fdf-4593-92e3-352f07e26062" path="/var/lib/kubelet/pods/2d48f238-7fdf-4593-92e3-352f07e26062/volumes" Dec 05 09:40:03 crc kubenswrapper[4815]: I1205 09:40:03.438625 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5de674b4-317a-4d40-959c-6e0b050a18de" path="/var/lib/kubelet/pods/5de674b4-317a-4d40-959c-6e0b050a18de/volumes" Dec 05 09:40:04 crc kubenswrapper[4815]: I1205 09:40:04.034013 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-qgqvs"] Dec 05 09:40:04 crc kubenswrapper[4815]: I1205 09:40:04.041163 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-d1d3-account-create-update-5gxdr"] Dec 05 09:40:04 crc kubenswrapper[4815]: I1205 09:40:04.048106 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-88f59"] Dec 05 09:40:04 crc kubenswrapper[4815]: I1205 09:40:04.057579 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-d1d3-account-create-update-5gxdr"] Dec 05 09:40:04 crc kubenswrapper[4815]: I1205 09:40:04.068402 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-qgqvs"] Dec 05 09:40:04 crc kubenswrapper[4815]: I1205 09:40:04.075778 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-88f59"] Dec 05 09:40:04 crc kubenswrapper[4815]: I1205 09:40:04.082826 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-41a2-account-create-update-rxfbj"] Dec 05 09:40:04 crc kubenswrapper[4815]: I1205 09:40:04.089125 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-41a2-account-create-update-rxfbj"] Dec 05 09:40:05 crc kubenswrapper[4815]: I1205 09:40:05.431446 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15e7e2fa-915a-407e-a8cd-1a0225692722" path="/var/lib/kubelet/pods/15e7e2fa-915a-407e-a8cd-1a0225692722/volumes" Dec 05 09:40:05 crc kubenswrapper[4815]: I1205 09:40:05.432570 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c" path="/var/lib/kubelet/pods/b1fbb4f0-9d07-46f4-8336-29f17f0b1a7c/volumes" Dec 05 09:40:05 crc kubenswrapper[4815]: I1205 09:40:05.433128 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be43b898-0145-4ee3-aeb0-0e35acaa094d" path="/var/lib/kubelet/pods/be43b898-0145-4ee3-aeb0-0e35acaa094d/volumes" Dec 05 09:40:05 crc kubenswrapper[4815]: I1205 09:40:05.433676 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0186806-095c-4433-bb68-55732c995bed" path="/var/lib/kubelet/pods/e0186806-095c-4433-bb68-55732c995bed/volumes" Dec 05 09:40:12 crc kubenswrapper[4815]: I1205 09:40:12.746228 4815 generic.go:334] "Generic (PLEG): container finished" podID="bf14d613-f583-4916-9a10-07fe2de25e48" containerID="ff883880d2cd98b119cf093dcfe70e10b93acbe38cca5a0f6b807a841aa2903b" exitCode=0 Dec 05 09:40:12 crc kubenswrapper[4815]: I1205 09:40:12.746321 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" event={"ID":"bf14d613-f583-4916-9a10-07fe2de25e48","Type":"ContainerDied","Data":"ff883880d2cd98b119cf093dcfe70e10b93acbe38cca5a0f6b807a841aa2903b"} Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.209544 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.307393 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-inventory\") pod \"bf14d613-f583-4916-9a10-07fe2de25e48\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.307461 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-ssh-key\") pod \"bf14d613-f583-4916-9a10-07fe2de25e48\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.307617 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wwsd\" (UniqueName: \"kubernetes.io/projected/bf14d613-f583-4916-9a10-07fe2de25e48-kube-api-access-2wwsd\") pod \"bf14d613-f583-4916-9a10-07fe2de25e48\" (UID: \"bf14d613-f583-4916-9a10-07fe2de25e48\") " Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.315590 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf14d613-f583-4916-9a10-07fe2de25e48-kube-api-access-2wwsd" (OuterVolumeSpecName: "kube-api-access-2wwsd") pod "bf14d613-f583-4916-9a10-07fe2de25e48" (UID: "bf14d613-f583-4916-9a10-07fe2de25e48"). InnerVolumeSpecName "kube-api-access-2wwsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.335008 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bf14d613-f583-4916-9a10-07fe2de25e48" (UID: "bf14d613-f583-4916-9a10-07fe2de25e48"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.345504 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-inventory" (OuterVolumeSpecName: "inventory") pod "bf14d613-f583-4916-9a10-07fe2de25e48" (UID: "bf14d613-f583-4916-9a10-07fe2de25e48"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.409868 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.409933 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bf14d613-f583-4916-9a10-07fe2de25e48-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.409945 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wwsd\" (UniqueName: \"kubernetes.io/projected/bf14d613-f583-4916-9a10-07fe2de25e48-kube-api-access-2wwsd\") on node \"crc\" DevicePath \"\"" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.770639 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" event={"ID":"bf14d613-f583-4916-9a10-07fe2de25e48","Type":"ContainerDied","Data":"0ddff0c6b7161d7a2a1bf5a1a2c9d17de9da3e24c4b9d7e8756cdb3c7b977e54"} Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.770684 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0ddff0c6b7161d7a2a1bf5a1a2c9d17de9da3e24c4b9d7e8756cdb3c7b977e54" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.770767 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.864471 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb"] Dec 05 09:40:14 crc kubenswrapper[4815]: E1205 09:40:14.870972 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf14d613-f583-4916-9a10-07fe2de25e48" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.871028 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf14d613-f583-4916-9a10-07fe2de25e48" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.871239 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf14d613-f583-4916-9a10-07fe2de25e48" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.871902 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.874259 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.874598 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.874745 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.883994 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb"] Dec 05 09:40:14 crc kubenswrapper[4815]: I1205 09:40:14.885093 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.019841 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.020159 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.020332 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km6hf\" (UniqueName: \"kubernetes.io/projected/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-kube-api-access-km6hf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.121818 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.121898 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km6hf\" (UniqueName: \"kubernetes.io/projected/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-kube-api-access-km6hf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.122116 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.127704 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.128402 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.142423 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km6hf\" (UniqueName: \"kubernetes.io/projected/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-kube-api-access-km6hf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.188116 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.735947 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb"] Dec 05 09:40:15 crc kubenswrapper[4815]: W1205 09:40:15.741786 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fd95c26_42fe_4d0c_82a0_eca943ceaf79.slice/crio-e1c952358e6dad16899f0814e270156553b394009aede903cb2f5c07c20816e8 WatchSource:0}: Error finding container e1c952358e6dad16899f0814e270156553b394009aede903cb2f5c07c20816e8: Status 404 returned error can't find the container with id e1c952358e6dad16899f0814e270156553b394009aede903cb2f5c07c20816e8 Dec 05 09:40:15 crc kubenswrapper[4815]: I1205 09:40:15.791827 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" event={"ID":"8fd95c26-42fe-4d0c-82a0-eca943ceaf79","Type":"ContainerStarted","Data":"e1c952358e6dad16899f0814e270156553b394009aede903cb2f5c07c20816e8"} Dec 05 09:40:16 crc kubenswrapper[4815]: I1205 09:40:16.801127 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" event={"ID":"8fd95c26-42fe-4d0c-82a0-eca943ceaf79","Type":"ContainerStarted","Data":"7af5b7a27fd38a7e4b8639ab8e6661d8f65c768ce29bc1dce549957fdbcfe4c7"} Dec 05 09:40:16 crc kubenswrapper[4815]: I1205 09:40:16.817854 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" podStartSLOduration=2.239340737 podStartE2EDuration="2.817835561s" podCreationTimestamp="2025-12-05 09:40:14 +0000 UTC" firstStartedPulling="2025-12-05 09:40:15.74764553 +0000 UTC m=+2014.626252377" lastFinishedPulling="2025-12-05 09:40:16.326140364 +0000 UTC m=+2015.204747201" observedRunningTime="2025-12-05 09:40:16.814389808 +0000 UTC m=+2015.692996645" watchObservedRunningTime="2025-12-05 09:40:16.817835561 +0000 UTC m=+2015.696442398" Dec 05 09:40:20 crc kubenswrapper[4815]: I1205 09:40:20.841954 4815 generic.go:334] "Generic (PLEG): container finished" podID="8fd95c26-42fe-4d0c-82a0-eca943ceaf79" containerID="7af5b7a27fd38a7e4b8639ab8e6661d8f65c768ce29bc1dce549957fdbcfe4c7" exitCode=0 Dec 05 09:40:20 crc kubenswrapper[4815]: I1205 09:40:20.842040 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" event={"ID":"8fd95c26-42fe-4d0c-82a0-eca943ceaf79","Type":"ContainerDied","Data":"7af5b7a27fd38a7e4b8639ab8e6661d8f65c768ce29bc1dce549957fdbcfe4c7"} Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.246983 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.372160 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km6hf\" (UniqueName: \"kubernetes.io/projected/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-kube-api-access-km6hf\") pod \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.372285 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-inventory\") pod \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.372307 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-ssh-key\") pod \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\" (UID: \"8fd95c26-42fe-4d0c-82a0-eca943ceaf79\") " Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.378161 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-kube-api-access-km6hf" (OuterVolumeSpecName: "kube-api-access-km6hf") pod "8fd95c26-42fe-4d0c-82a0-eca943ceaf79" (UID: "8fd95c26-42fe-4d0c-82a0-eca943ceaf79"). InnerVolumeSpecName "kube-api-access-km6hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.398638 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-inventory" (OuterVolumeSpecName: "inventory") pod "8fd95c26-42fe-4d0c-82a0-eca943ceaf79" (UID: "8fd95c26-42fe-4d0c-82a0-eca943ceaf79"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.418967 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8fd95c26-42fe-4d0c-82a0-eca943ceaf79" (UID: "8fd95c26-42fe-4d0c-82a0-eca943ceaf79"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.474685 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km6hf\" (UniqueName: \"kubernetes.io/projected/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-kube-api-access-km6hf\") on node \"crc\" DevicePath \"\"" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.474719 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.474730 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8fd95c26-42fe-4d0c-82a0-eca943ceaf79-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.861630 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" event={"ID":"8fd95c26-42fe-4d0c-82a0-eca943ceaf79","Type":"ContainerDied","Data":"e1c952358e6dad16899f0814e270156553b394009aede903cb2f5c07c20816e8"} Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.861677 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1c952358e6dad16899f0814e270156553b394009aede903cb2f5c07c20816e8" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.861689 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.952827 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd"] Dec 05 09:40:22 crc kubenswrapper[4815]: E1205 09:40:22.953365 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fd95c26-42fe-4d0c-82a0-eca943ceaf79" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.953395 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fd95c26-42fe-4d0c-82a0-eca943ceaf79" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.953707 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fd95c26-42fe-4d0c-82a0-eca943ceaf79" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.954469 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.965154 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.965462 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.965532 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.965615 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.969637 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd"] Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.988964 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72l9v\" (UniqueName: \"kubernetes.io/projected/c84f99f1-df65-4728-a42b-6b9d5658a228-kube-api-access-72l9v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.989091 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:40:22 crc kubenswrapper[4815]: I1205 09:40:22.989206 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:40:23 crc kubenswrapper[4815]: I1205 09:40:23.091270 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72l9v\" (UniqueName: \"kubernetes.io/projected/c84f99f1-df65-4728-a42b-6b9d5658a228-kube-api-access-72l9v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:40:23 crc kubenswrapper[4815]: I1205 09:40:23.091565 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:40:23 crc kubenswrapper[4815]: I1205 09:40:23.091656 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:40:23 crc kubenswrapper[4815]: I1205 09:40:23.096371 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:40:23 crc kubenswrapper[4815]: I1205 09:40:23.098900 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:40:23 crc kubenswrapper[4815]: I1205 09:40:23.106131 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72l9v\" (UniqueName: \"kubernetes.io/projected/c84f99f1-df65-4728-a42b-6b9d5658a228-kube-api-access-72l9v\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:40:23 crc kubenswrapper[4815]: I1205 09:40:23.290366 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:40:23 crc kubenswrapper[4815]: I1205 09:40:23.828313 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd"] Dec 05 09:40:23 crc kubenswrapper[4815]: W1205 09:40:23.851792 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc84f99f1_df65_4728_a42b_6b9d5658a228.slice/crio-c9ae23635fc078ad3c4016f053f58a848886f072b97568904597766dfb535611 WatchSource:0}: Error finding container c9ae23635fc078ad3c4016f053f58a848886f072b97568904597766dfb535611: Status 404 returned error can't find the container with id c9ae23635fc078ad3c4016f053f58a848886f072b97568904597766dfb535611 Dec 05 09:40:23 crc kubenswrapper[4815]: I1205 09:40:23.875245 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" event={"ID":"c84f99f1-df65-4728-a42b-6b9d5658a228","Type":"ContainerStarted","Data":"c9ae23635fc078ad3c4016f053f58a848886f072b97568904597766dfb535611"} Dec 05 09:40:24 crc kubenswrapper[4815]: I1205 09:40:24.884219 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" event={"ID":"c84f99f1-df65-4728-a42b-6b9d5658a228","Type":"ContainerStarted","Data":"a20a9e53498757533027b60b62550f6e5fe588334703059cc6248532d860f214"} Dec 05 09:40:24 crc kubenswrapper[4815]: I1205 09:40:24.904323 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" podStartSLOduration=2.408044741 podStartE2EDuration="2.904305622s" podCreationTimestamp="2025-12-05 09:40:22 +0000 UTC" firstStartedPulling="2025-12-05 09:40:23.856101585 +0000 UTC m=+2022.734708422" lastFinishedPulling="2025-12-05 09:40:24.352362456 +0000 UTC m=+2023.230969303" observedRunningTime="2025-12-05 09:40:24.901511087 +0000 UTC m=+2023.780117934" watchObservedRunningTime="2025-12-05 09:40:24.904305622 +0000 UTC m=+2023.782912459" Dec 05 09:40:36 crc kubenswrapper[4815]: I1205 09:40:36.050944 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s5qnf"] Dec 05 09:40:36 crc kubenswrapper[4815]: I1205 09:40:36.058269 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-s5qnf"] Dec 05 09:40:37 crc kubenswrapper[4815]: I1205 09:40:37.431054 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9371c4b-130e-4c75-b54a-7dc4d07fe7e4" path="/var/lib/kubelet/pods/a9371c4b-130e-4c75-b54a-7dc4d07fe7e4/volumes" Dec 05 09:40:50 crc kubenswrapper[4815]: I1205 09:40:50.192350 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:40:50 crc kubenswrapper[4815]: I1205 09:40:50.192925 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:40:58 crc kubenswrapper[4815]: I1205 09:40:58.555854 4815 scope.go:117] "RemoveContainer" containerID="71d7348eefb9baa02cf9473eb142136a4611deeb4abd2da64031b8f3fc0c03fe" Dec 05 09:40:58 crc kubenswrapper[4815]: I1205 09:40:58.598614 4815 scope.go:117] "RemoveContainer" containerID="61824a91f472cb2f73bb1877b3635263a8793a54054fe0745a6ea6af0604561a" Dec 05 09:40:58 crc kubenswrapper[4815]: I1205 09:40:58.635551 4815 scope.go:117] "RemoveContainer" containerID="46f2696aa70c8ba1f1e30cd56782981a9f0fbdffd5d57038b0d543109e345293" Dec 05 09:40:58 crc kubenswrapper[4815]: I1205 09:40:58.718062 4815 scope.go:117] "RemoveContainer" containerID="158b471d9b8af04deba859b22bffd924e33ea5f08b023a3d7328196857b8eacf" Dec 05 09:40:58 crc kubenswrapper[4815]: I1205 09:40:58.740724 4815 scope.go:117] "RemoveContainer" containerID="5959f1dc92563e332c1be778e899ecdc1f1000da1aec244ee4903d2aad1ce66d" Dec 05 09:40:58 crc kubenswrapper[4815]: I1205 09:40:58.782662 4815 scope.go:117] "RemoveContainer" containerID="ffe9faeaa0bf8cd09661f677d909031160c5256658cc3eb14e3506e73ac4bd46" Dec 05 09:40:58 crc kubenswrapper[4815]: I1205 09:40:58.855092 4815 scope.go:117] "RemoveContainer" containerID="e7948e689586307da965d0d523ff5693636db131adf753f832413cf37fcde6d8" Dec 05 09:41:00 crc kubenswrapper[4815]: I1205 09:41:00.035986 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-jxrzv"] Dec 05 09:41:00 crc kubenswrapper[4815]: I1205 09:41:00.045626 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-jxrzv"] Dec 05 09:41:00 crc kubenswrapper[4815]: I1205 09:41:00.054761 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dxtfw"] Dec 05 09:41:00 crc kubenswrapper[4815]: I1205 09:41:00.065742 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dxtfw"] Dec 05 09:41:01 crc kubenswrapper[4815]: I1205 09:41:01.434058 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04a96446-9720-4d97-a657-42ab3826aee3" path="/var/lib/kubelet/pods/04a96446-9720-4d97-a657-42ab3826aee3/volumes" Dec 05 09:41:01 crc kubenswrapper[4815]: I1205 09:41:01.434752 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f42cc49-79a7-494e-9f3e-da2da9da9940" path="/var/lib/kubelet/pods/6f42cc49-79a7-494e-9f3e-da2da9da9940/volumes" Dec 05 09:41:19 crc kubenswrapper[4815]: I1205 09:41:19.394290 4815 generic.go:334] "Generic (PLEG): container finished" podID="c84f99f1-df65-4728-a42b-6b9d5658a228" containerID="a20a9e53498757533027b60b62550f6e5fe588334703059cc6248532d860f214" exitCode=0 Dec 05 09:41:19 crc kubenswrapper[4815]: I1205 09:41:19.394377 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" event={"ID":"c84f99f1-df65-4728-a42b-6b9d5658a228","Type":"ContainerDied","Data":"a20a9e53498757533027b60b62550f6e5fe588334703059cc6248532d860f214"} Dec 05 09:41:20 crc kubenswrapper[4815]: I1205 09:41:20.192545 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:41:20 crc kubenswrapper[4815]: I1205 09:41:20.192644 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:41:20 crc kubenswrapper[4815]: I1205 09:41:20.897036 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.041599 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72l9v\" (UniqueName: \"kubernetes.io/projected/c84f99f1-df65-4728-a42b-6b9d5658a228-kube-api-access-72l9v\") pod \"c84f99f1-df65-4728-a42b-6b9d5658a228\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.041675 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-inventory\") pod \"c84f99f1-df65-4728-a42b-6b9d5658a228\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.041804 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-ssh-key\") pod \"c84f99f1-df65-4728-a42b-6b9d5658a228\" (UID: \"c84f99f1-df65-4728-a42b-6b9d5658a228\") " Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.056884 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c84f99f1-df65-4728-a42b-6b9d5658a228-kube-api-access-72l9v" (OuterVolumeSpecName: "kube-api-access-72l9v") pod "c84f99f1-df65-4728-a42b-6b9d5658a228" (UID: "c84f99f1-df65-4728-a42b-6b9d5658a228"). InnerVolumeSpecName "kube-api-access-72l9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.069620 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c84f99f1-df65-4728-a42b-6b9d5658a228" (UID: "c84f99f1-df65-4728-a42b-6b9d5658a228"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.075476 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-inventory" (OuterVolumeSpecName: "inventory") pod "c84f99f1-df65-4728-a42b-6b9d5658a228" (UID: "c84f99f1-df65-4728-a42b-6b9d5658a228"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.144469 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72l9v\" (UniqueName: \"kubernetes.io/projected/c84f99f1-df65-4728-a42b-6b9d5658a228-kube-api-access-72l9v\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.144528 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.144539 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c84f99f1-df65-4728-a42b-6b9d5658a228-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.418724 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.439507 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd" event={"ID":"c84f99f1-df65-4728-a42b-6b9d5658a228","Type":"ContainerDied","Data":"c9ae23635fc078ad3c4016f053f58a848886f072b97568904597766dfb535611"} Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.439561 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9ae23635fc078ad3c4016f053f58a848886f072b97568904597766dfb535611" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.537676 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-t7d9z"] Dec 05 09:41:21 crc kubenswrapper[4815]: E1205 09:41:21.538133 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c84f99f1-df65-4728-a42b-6b9d5658a228" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.538156 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c84f99f1-df65-4728-a42b-6b9d5658a228" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.538360 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c84f99f1-df65-4728-a42b-6b9d5658a228" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.539027 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.543719 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.544000 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.544120 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.545473 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.558258 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grds8\" (UniqueName: \"kubernetes.io/projected/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-kube-api-access-grds8\") pod \"ssh-known-hosts-edpm-deployment-t7d9z\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.558748 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-t7d9z\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.558961 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-t7d9z\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.566803 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-t7d9z"] Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.659798 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grds8\" (UniqueName: \"kubernetes.io/projected/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-kube-api-access-grds8\") pod \"ssh-known-hosts-edpm-deployment-t7d9z\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.660198 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-t7d9z\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.660235 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-t7d9z\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.665181 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-t7d9z\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.666959 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-t7d9z\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.678659 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grds8\" (UniqueName: \"kubernetes.io/projected/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-kube-api-access-grds8\") pod \"ssh-known-hosts-edpm-deployment-t7d9z\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:21 crc kubenswrapper[4815]: I1205 09:41:21.867024 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:22 crc kubenswrapper[4815]: I1205 09:41:22.438060 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-t7d9z"] Dec 05 09:41:23 crc kubenswrapper[4815]: I1205 09:41:23.446879 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" event={"ID":"83d6ef02-36e2-49bc-a1cd-5225686a0fdd","Type":"ContainerStarted","Data":"9402a2252e503ec8bb7473f51f6206e13d02721eb4b4c0f25114c18ba020b380"} Dec 05 09:41:23 crc kubenswrapper[4815]: I1205 09:41:23.448118 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" event={"ID":"83d6ef02-36e2-49bc-a1cd-5225686a0fdd","Type":"ContainerStarted","Data":"03dca5175513bce36966d2122db74ba01499733f780752b555eec6dc9a90d246"} Dec 05 09:41:32 crc kubenswrapper[4815]: I1205 09:41:32.517260 4815 generic.go:334] "Generic (PLEG): container finished" podID="83d6ef02-36e2-49bc-a1cd-5225686a0fdd" containerID="9402a2252e503ec8bb7473f51f6206e13d02721eb4b4c0f25114c18ba020b380" exitCode=0 Dec 05 09:41:32 crc kubenswrapper[4815]: I1205 09:41:32.517847 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" event={"ID":"83d6ef02-36e2-49bc-a1cd-5225686a0fdd","Type":"ContainerDied","Data":"9402a2252e503ec8bb7473f51f6206e13d02721eb4b4c0f25114c18ba020b380"} Dec 05 09:41:33 crc kubenswrapper[4815]: I1205 09:41:33.912106 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:33 crc kubenswrapper[4815]: I1205 09:41:33.968755 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-inventory-0\") pod \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " Dec 05 09:41:33 crc kubenswrapper[4815]: I1205 09:41:33.968804 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-ssh-key-openstack-edpm-ipam\") pod \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " Dec 05 09:41:33 crc kubenswrapper[4815]: I1205 09:41:33.968831 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grds8\" (UniqueName: \"kubernetes.io/projected/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-kube-api-access-grds8\") pod \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\" (UID: \"83d6ef02-36e2-49bc-a1cd-5225686a0fdd\") " Dec 05 09:41:33 crc kubenswrapper[4815]: I1205 09:41:33.973960 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-kube-api-access-grds8" (OuterVolumeSpecName: "kube-api-access-grds8") pod "83d6ef02-36e2-49bc-a1cd-5225686a0fdd" (UID: "83d6ef02-36e2-49bc-a1cd-5225686a0fdd"). InnerVolumeSpecName "kube-api-access-grds8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:41:33 crc kubenswrapper[4815]: I1205 09:41:33.995330 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "83d6ef02-36e2-49bc-a1cd-5225686a0fdd" (UID: "83d6ef02-36e2-49bc-a1cd-5225686a0fdd"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:41:33 crc kubenswrapper[4815]: I1205 09:41:33.997452 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "83d6ef02-36e2-49bc-a1cd-5225686a0fdd" (UID: "83d6ef02-36e2-49bc-a1cd-5225686a0fdd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.070848 4815 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.070897 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.070924 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grds8\" (UniqueName: \"kubernetes.io/projected/83d6ef02-36e2-49bc-a1cd-5225686a0fdd-kube-api-access-grds8\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.534726 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" event={"ID":"83d6ef02-36e2-49bc-a1cd-5225686a0fdd","Type":"ContainerDied","Data":"03dca5175513bce36966d2122db74ba01499733f780752b555eec6dc9a90d246"} Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.534772 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="03dca5175513bce36966d2122db74ba01499733f780752b555eec6dc9a90d246" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.534849 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-t7d9z" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.626556 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6"] Dec 05 09:41:34 crc kubenswrapper[4815]: E1205 09:41:34.627012 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d6ef02-36e2-49bc-a1cd-5225686a0fdd" containerName="ssh-known-hosts-edpm-deployment" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.627033 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d6ef02-36e2-49bc-a1cd-5225686a0fdd" containerName="ssh-known-hosts-edpm-deployment" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.627222 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d6ef02-36e2-49bc-a1cd-5225686a0fdd" containerName="ssh-known-hosts-edpm-deployment" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.628863 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.631799 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.631961 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.632191 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.638942 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6"] Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.643040 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.682257 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9pq6\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.682652 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9pq6\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.682719 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz4xt\" (UniqueName: \"kubernetes.io/projected/6a49a6d7-0038-47ab-acea-549233b7d78f-kube-api-access-vz4xt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9pq6\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.784942 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9pq6\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.785007 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz4xt\" (UniqueName: \"kubernetes.io/projected/6a49a6d7-0038-47ab-acea-549233b7d78f-kube-api-access-vz4xt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9pq6\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.785031 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9pq6\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.789352 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9pq6\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.790320 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9pq6\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.821190 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz4xt\" (UniqueName: \"kubernetes.io/projected/6a49a6d7-0038-47ab-acea-549233b7d78f-kube-api-access-vz4xt\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-j9pq6\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:34 crc kubenswrapper[4815]: I1205 09:41:34.946981 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:35 crc kubenswrapper[4815]: I1205 09:41:35.479535 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6"] Dec 05 09:41:35 crc kubenswrapper[4815]: I1205 09:41:35.545417 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" event={"ID":"6a49a6d7-0038-47ab-acea-549233b7d78f","Type":"ContainerStarted","Data":"6666b67ece40886c4406f9237778d746aef2692f257b85d68cbda25f18f2f419"} Dec 05 09:41:36 crc kubenswrapper[4815]: I1205 09:41:36.555309 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" event={"ID":"6a49a6d7-0038-47ab-acea-549233b7d78f","Type":"ContainerStarted","Data":"ea298044d76fcd6d85f472282a73221c7b63dbbc05cb96ad92241cf7cbf54eee"} Dec 05 09:41:36 crc kubenswrapper[4815]: I1205 09:41:36.582113 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" podStartSLOduration=1.835259328 podStartE2EDuration="2.582081394s" podCreationTimestamp="2025-12-05 09:41:34 +0000 UTC" firstStartedPulling="2025-12-05 09:41:35.490902614 +0000 UTC m=+2094.369509451" lastFinishedPulling="2025-12-05 09:41:36.23772468 +0000 UTC m=+2095.116331517" observedRunningTime="2025-12-05 09:41:36.578143838 +0000 UTC m=+2095.456750665" watchObservedRunningTime="2025-12-05 09:41:36.582081394 +0000 UTC m=+2095.460688231" Dec 05 09:41:43 crc kubenswrapper[4815]: I1205 09:41:43.049557 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-2k4k5"] Dec 05 09:41:43 crc kubenswrapper[4815]: I1205 09:41:43.056383 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-2k4k5"] Dec 05 09:41:43 crc kubenswrapper[4815]: I1205 09:41:43.428990 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8677ce04-f49f-4df4-ad0f-ee434e75799b" path="/var/lib/kubelet/pods/8677ce04-f49f-4df4-ad0f-ee434e75799b/volumes" Dec 05 09:41:46 crc kubenswrapper[4815]: I1205 09:41:46.642028 4815 generic.go:334] "Generic (PLEG): container finished" podID="6a49a6d7-0038-47ab-acea-549233b7d78f" containerID="ea298044d76fcd6d85f472282a73221c7b63dbbc05cb96ad92241cf7cbf54eee" exitCode=0 Dec 05 09:41:46 crc kubenswrapper[4815]: I1205 09:41:46.642114 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" event={"ID":"6a49a6d7-0038-47ab-acea-549233b7d78f","Type":"ContainerDied","Data":"ea298044d76fcd6d85f472282a73221c7b63dbbc05cb96ad92241cf7cbf54eee"} Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.063794 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.256282 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-ssh-key\") pod \"6a49a6d7-0038-47ab-acea-549233b7d78f\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.256335 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vz4xt\" (UniqueName: \"kubernetes.io/projected/6a49a6d7-0038-47ab-acea-549233b7d78f-kube-api-access-vz4xt\") pod \"6a49a6d7-0038-47ab-acea-549233b7d78f\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.256355 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-inventory\") pod \"6a49a6d7-0038-47ab-acea-549233b7d78f\" (UID: \"6a49a6d7-0038-47ab-acea-549233b7d78f\") " Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.274777 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a49a6d7-0038-47ab-acea-549233b7d78f-kube-api-access-vz4xt" (OuterVolumeSpecName: "kube-api-access-vz4xt") pod "6a49a6d7-0038-47ab-acea-549233b7d78f" (UID: "6a49a6d7-0038-47ab-acea-549233b7d78f"). InnerVolumeSpecName "kube-api-access-vz4xt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.291616 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-inventory" (OuterVolumeSpecName: "inventory") pod "6a49a6d7-0038-47ab-acea-549233b7d78f" (UID: "6a49a6d7-0038-47ab-acea-549233b7d78f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.335348 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6a49a6d7-0038-47ab-acea-549233b7d78f" (UID: "6a49a6d7-0038-47ab-acea-549233b7d78f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.358142 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.358424 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vz4xt\" (UniqueName: \"kubernetes.io/projected/6a49a6d7-0038-47ab-acea-549233b7d78f-kube-api-access-vz4xt\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.358808 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6a49a6d7-0038-47ab-acea-549233b7d78f-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.659116 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" event={"ID":"6a49a6d7-0038-47ab-acea-549233b7d78f","Type":"ContainerDied","Data":"6666b67ece40886c4406f9237778d746aef2692f257b85d68cbda25f18f2f419"} Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.659377 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6666b67ece40886c4406f9237778d746aef2692f257b85d68cbda25f18f2f419" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.659200 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.759002 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts"] Dec 05 09:41:48 crc kubenswrapper[4815]: E1205 09:41:48.759428 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a49a6d7-0038-47ab-acea-549233b7d78f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.759450 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a49a6d7-0038-47ab-acea-549233b7d78f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.759734 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a49a6d7-0038-47ab-acea-549233b7d78f" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.760442 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.767246 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.767472 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.767650 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.767822 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.770116 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts"] Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.871817 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qthts\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.872223 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv5zv\" (UniqueName: \"kubernetes.io/projected/f29fa7d7-8396-42f6-b244-ad560a622193-kube-api-access-jv5zv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qthts\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.872290 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qthts\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.974127 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qthts\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.974381 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv5zv\" (UniqueName: \"kubernetes.io/projected/f29fa7d7-8396-42f6-b244-ad560a622193-kube-api-access-jv5zv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qthts\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.974510 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qthts\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.978268 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qthts\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.986425 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qthts\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:41:48 crc kubenswrapper[4815]: I1205 09:41:48.996902 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv5zv\" (UniqueName: \"kubernetes.io/projected/f29fa7d7-8396-42f6-b244-ad560a622193-kube-api-access-jv5zv\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-qthts\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:41:49 crc kubenswrapper[4815]: I1205 09:41:49.084558 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:41:49 crc kubenswrapper[4815]: W1205 09:41:49.421507 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf29fa7d7_8396_42f6_b244_ad560a622193.slice/crio-e45dc8d501a5cf22e7fe4ce08785fa35d1ea2495b266f8e20774901d893e5a55 WatchSource:0}: Error finding container e45dc8d501a5cf22e7fe4ce08785fa35d1ea2495b266f8e20774901d893e5a55: Status 404 returned error can't find the container with id e45dc8d501a5cf22e7fe4ce08785fa35d1ea2495b266f8e20774901d893e5a55 Dec 05 09:41:49 crc kubenswrapper[4815]: I1205 09:41:49.429801 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts"] Dec 05 09:41:49 crc kubenswrapper[4815]: I1205 09:41:49.666762 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" event={"ID":"f29fa7d7-8396-42f6-b244-ad560a622193","Type":"ContainerStarted","Data":"e45dc8d501a5cf22e7fe4ce08785fa35d1ea2495b266f8e20774901d893e5a55"} Dec 05 09:41:50 crc kubenswrapper[4815]: I1205 09:41:50.192755 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:41:50 crc kubenswrapper[4815]: I1205 09:41:50.194080 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:41:50 crc kubenswrapper[4815]: I1205 09:41:50.194227 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:41:50 crc kubenswrapper[4815]: I1205 09:41:50.195180 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1db619e28f79e22d68fc8699b12d64acf5e1c448e650971b8846c7b1098b9c14"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:41:50 crc kubenswrapper[4815]: I1205 09:41:50.195307 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://1db619e28f79e22d68fc8699b12d64acf5e1c448e650971b8846c7b1098b9c14" gracePeriod=600 Dec 05 09:41:50 crc kubenswrapper[4815]: I1205 09:41:50.674114 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" event={"ID":"f29fa7d7-8396-42f6-b244-ad560a622193","Type":"ContainerStarted","Data":"8a9e01c8f8d2a31a2064be85e6cc1a27e715db83bea23b7197664bedc70f37a6"} Dec 05 09:41:50 crc kubenswrapper[4815]: I1205 09:41:50.677444 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="1db619e28f79e22d68fc8699b12d64acf5e1c448e650971b8846c7b1098b9c14" exitCode=0 Dec 05 09:41:50 crc kubenswrapper[4815]: I1205 09:41:50.677484 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"1db619e28f79e22d68fc8699b12d64acf5e1c448e650971b8846c7b1098b9c14"} Dec 05 09:41:50 crc kubenswrapper[4815]: I1205 09:41:50.677529 4815 scope.go:117] "RemoveContainer" containerID="c41f3f1db2734332e7f0685b4590335a05b5a05ba2ddb887151defcb0b74adfa" Dec 05 09:41:52 crc kubenswrapper[4815]: I1205 09:41:52.701820 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372"} Dec 05 09:41:52 crc kubenswrapper[4815]: I1205 09:41:52.721780 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" podStartSLOduration=4.319052102 podStartE2EDuration="4.721760781s" podCreationTimestamp="2025-12-05 09:41:48 +0000 UTC" firstStartedPulling="2025-12-05 09:41:49.423761198 +0000 UTC m=+2108.302368035" lastFinishedPulling="2025-12-05 09:41:49.826469877 +0000 UTC m=+2108.705076714" observedRunningTime="2025-12-05 09:41:50.698033031 +0000 UTC m=+2109.576639868" watchObservedRunningTime="2025-12-05 09:41:52.721760781 +0000 UTC m=+2111.600367618" Dec 05 09:41:59 crc kubenswrapper[4815]: I1205 09:41:59.015027 4815 scope.go:117] "RemoveContainer" containerID="f5127d3ec099ece6f2dc2b9875ac7d19cfe4af3a2626d723607dcf76d07b7498" Dec 05 09:41:59 crc kubenswrapper[4815]: I1205 09:41:59.074279 4815 scope.go:117] "RemoveContainer" containerID="41ea05b3ae7a09bbe7565edab94ed72b396859a198682ebbcb08d99b0730e0ba" Dec 05 09:41:59 crc kubenswrapper[4815]: I1205 09:41:59.127466 4815 scope.go:117] "RemoveContainer" containerID="469428e29fd165fd4b6fa5f86a059013f338dabaa038f8df298eca9291560173" Dec 05 09:42:01 crc kubenswrapper[4815]: I1205 09:42:01.790428 4815 generic.go:334] "Generic (PLEG): container finished" podID="f29fa7d7-8396-42f6-b244-ad560a622193" containerID="8a9e01c8f8d2a31a2064be85e6cc1a27e715db83bea23b7197664bedc70f37a6" exitCode=0 Dec 05 09:42:01 crc kubenswrapper[4815]: I1205 09:42:01.791098 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" event={"ID":"f29fa7d7-8396-42f6-b244-ad560a622193","Type":"ContainerDied","Data":"8a9e01c8f8d2a31a2064be85e6cc1a27e715db83bea23b7197664bedc70f37a6"} Dec 05 09:42:03 crc kubenswrapper[4815]: I1205 09:42:03.827694 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" event={"ID":"f29fa7d7-8396-42f6-b244-ad560a622193","Type":"ContainerDied","Data":"e45dc8d501a5cf22e7fe4ce08785fa35d1ea2495b266f8e20774901d893e5a55"} Dec 05 09:42:03 crc kubenswrapper[4815]: I1205 09:42:03.827956 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e45dc8d501a5cf22e7fe4ce08785fa35d1ea2495b266f8e20774901d893e5a55" Dec 05 09:42:03 crc kubenswrapper[4815]: I1205 09:42:03.874393 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:42:03 crc kubenswrapper[4815]: I1205 09:42:03.972037 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-inventory\") pod \"f29fa7d7-8396-42f6-b244-ad560a622193\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " Dec 05 09:42:03 crc kubenswrapper[4815]: I1205 09:42:03.972139 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jv5zv\" (UniqueName: \"kubernetes.io/projected/f29fa7d7-8396-42f6-b244-ad560a622193-kube-api-access-jv5zv\") pod \"f29fa7d7-8396-42f6-b244-ad560a622193\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " Dec 05 09:42:03 crc kubenswrapper[4815]: I1205 09:42:03.972251 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-ssh-key\") pod \"f29fa7d7-8396-42f6-b244-ad560a622193\" (UID: \"f29fa7d7-8396-42f6-b244-ad560a622193\") " Dec 05 09:42:03 crc kubenswrapper[4815]: I1205 09:42:03.978709 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f29fa7d7-8396-42f6-b244-ad560a622193-kube-api-access-jv5zv" (OuterVolumeSpecName: "kube-api-access-jv5zv") pod "f29fa7d7-8396-42f6-b244-ad560a622193" (UID: "f29fa7d7-8396-42f6-b244-ad560a622193"). InnerVolumeSpecName "kube-api-access-jv5zv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:42:03 crc kubenswrapper[4815]: I1205 09:42:03.997809 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f29fa7d7-8396-42f6-b244-ad560a622193" (UID: "f29fa7d7-8396-42f6-b244-ad560a622193"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:42:03 crc kubenswrapper[4815]: I1205 09:42:03.998180 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-inventory" (OuterVolumeSpecName: "inventory") pod "f29fa7d7-8396-42f6-b244-ad560a622193" (UID: "f29fa7d7-8396-42f6-b244-ad560a622193"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:42:04 crc kubenswrapper[4815]: I1205 09:42:04.074642 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:42:04 crc kubenswrapper[4815]: I1205 09:42:04.074934 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jv5zv\" (UniqueName: \"kubernetes.io/projected/f29fa7d7-8396-42f6-b244-ad560a622193-kube-api-access-jv5zv\") on node \"crc\" DevicePath \"\"" Dec 05 09:42:04 crc kubenswrapper[4815]: I1205 09:42:04.074945 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f29fa7d7-8396-42f6-b244-ad560a622193-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:42:04 crc kubenswrapper[4815]: I1205 09:42:04.836430 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.239574 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9mkhb"] Dec 05 09:42:32 crc kubenswrapper[4815]: E1205 09:42:32.240581 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f29fa7d7-8396-42f6-b244-ad560a622193" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.240603 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="f29fa7d7-8396-42f6-b244-ad560a622193" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.240854 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="f29fa7d7-8396-42f6-b244-ad560a622193" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.242629 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.254771 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mkhb"] Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.303792 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htcxc\" (UniqueName: \"kubernetes.io/projected/727c4472-c2ef-41fb-a337-a9ec52aba849-kube-api-access-htcxc\") pod \"redhat-marketplace-9mkhb\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.303831 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-utilities\") pod \"redhat-marketplace-9mkhb\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.303931 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-catalog-content\") pod \"redhat-marketplace-9mkhb\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.405153 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htcxc\" (UniqueName: \"kubernetes.io/projected/727c4472-c2ef-41fb-a337-a9ec52aba849-kube-api-access-htcxc\") pod \"redhat-marketplace-9mkhb\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.405604 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-utilities\") pod \"redhat-marketplace-9mkhb\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.406178 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-utilities\") pod \"redhat-marketplace-9mkhb\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.406341 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-catalog-content\") pod \"redhat-marketplace-9mkhb\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.406831 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-catalog-content\") pod \"redhat-marketplace-9mkhb\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.426576 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htcxc\" (UniqueName: \"kubernetes.io/projected/727c4472-c2ef-41fb-a337-a9ec52aba849-kube-api-access-htcxc\") pod \"redhat-marketplace-9mkhb\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:32 crc kubenswrapper[4815]: I1205 09:42:32.561414 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:33 crc kubenswrapper[4815]: I1205 09:42:33.128562 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mkhb"] Dec 05 09:42:34 crc kubenswrapper[4815]: I1205 09:42:34.083573 4815 generic.go:334] "Generic (PLEG): container finished" podID="727c4472-c2ef-41fb-a337-a9ec52aba849" containerID="b236d9ec888fa2714fc04a708cb6a35e20472f064dae0adf1d514f871d5f4e47" exitCode=0 Dec 05 09:42:34 crc kubenswrapper[4815]: I1205 09:42:34.083624 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mkhb" event={"ID":"727c4472-c2ef-41fb-a337-a9ec52aba849","Type":"ContainerDied","Data":"b236d9ec888fa2714fc04a708cb6a35e20472f064dae0adf1d514f871d5f4e47"} Dec 05 09:42:34 crc kubenswrapper[4815]: I1205 09:42:34.084082 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mkhb" event={"ID":"727c4472-c2ef-41fb-a337-a9ec52aba849","Type":"ContainerStarted","Data":"0157598ba1a4e8fd8393f9c8e62e758a228b410b3cbf864d9f75a5405f227971"} Dec 05 09:42:34 crc kubenswrapper[4815]: I1205 09:42:34.812457 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-b55qd"] Dec 05 09:42:34 crc kubenswrapper[4815]: I1205 09:42:34.815994 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:34 crc kubenswrapper[4815]: I1205 09:42:34.835284 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b55qd"] Dec 05 09:42:34 crc kubenswrapper[4815]: I1205 09:42:34.954737 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-utilities\") pod \"certified-operators-b55qd\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:34 crc kubenswrapper[4815]: I1205 09:42:34.954812 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-catalog-content\") pod \"certified-operators-b55qd\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:34 crc kubenswrapper[4815]: I1205 09:42:34.954954 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6hr5\" (UniqueName: \"kubernetes.io/projected/962a4d01-45df-47a9-a9da-b36db0314c2d-kube-api-access-z6hr5\") pod \"certified-operators-b55qd\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:35 crc kubenswrapper[4815]: I1205 09:42:35.056640 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6hr5\" (UniqueName: \"kubernetes.io/projected/962a4d01-45df-47a9-a9da-b36db0314c2d-kube-api-access-z6hr5\") pod \"certified-operators-b55qd\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:35 crc kubenswrapper[4815]: I1205 09:42:35.056805 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-utilities\") pod \"certified-operators-b55qd\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:35 crc kubenswrapper[4815]: I1205 09:42:35.056843 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-catalog-content\") pod \"certified-operators-b55qd\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:35 crc kubenswrapper[4815]: I1205 09:42:35.057270 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-utilities\") pod \"certified-operators-b55qd\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:35 crc kubenswrapper[4815]: I1205 09:42:35.057303 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-catalog-content\") pod \"certified-operators-b55qd\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:35 crc kubenswrapper[4815]: I1205 09:42:35.086831 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6hr5\" (UniqueName: \"kubernetes.io/projected/962a4d01-45df-47a9-a9da-b36db0314c2d-kube-api-access-z6hr5\") pod \"certified-operators-b55qd\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:35 crc kubenswrapper[4815]: I1205 09:42:35.095050 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mkhb" event={"ID":"727c4472-c2ef-41fb-a337-a9ec52aba849","Type":"ContainerStarted","Data":"5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee"} Dec 05 09:42:35 crc kubenswrapper[4815]: I1205 09:42:35.142956 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:35 crc kubenswrapper[4815]: I1205 09:42:35.641124 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-b55qd"] Dec 05 09:42:36 crc kubenswrapper[4815]: I1205 09:42:36.103646 4815 generic.go:334] "Generic (PLEG): container finished" podID="962a4d01-45df-47a9-a9da-b36db0314c2d" containerID="a40ca4d2c88f26dd2a7290cde852a124a7cf45063ee99e451f8da186df77ddee" exitCode=0 Dec 05 09:42:36 crc kubenswrapper[4815]: I1205 09:42:36.103968 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b55qd" event={"ID":"962a4d01-45df-47a9-a9da-b36db0314c2d","Type":"ContainerDied","Data":"a40ca4d2c88f26dd2a7290cde852a124a7cf45063ee99e451f8da186df77ddee"} Dec 05 09:42:36 crc kubenswrapper[4815]: I1205 09:42:36.103995 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b55qd" event={"ID":"962a4d01-45df-47a9-a9da-b36db0314c2d","Type":"ContainerStarted","Data":"4c3db9748ebecf1546d178476294c5fe2cfe8be3d56416ce36fb96e16bae6d9a"} Dec 05 09:42:36 crc kubenswrapper[4815]: I1205 09:42:36.107513 4815 generic.go:334] "Generic (PLEG): container finished" podID="727c4472-c2ef-41fb-a337-a9ec52aba849" containerID="5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee" exitCode=0 Dec 05 09:42:36 crc kubenswrapper[4815]: I1205 09:42:36.107541 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mkhb" event={"ID":"727c4472-c2ef-41fb-a337-a9ec52aba849","Type":"ContainerDied","Data":"5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee"} Dec 05 09:42:37 crc kubenswrapper[4815]: I1205 09:42:37.118891 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b55qd" event={"ID":"962a4d01-45df-47a9-a9da-b36db0314c2d","Type":"ContainerStarted","Data":"e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397"} Dec 05 09:42:37 crc kubenswrapper[4815]: I1205 09:42:37.123001 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mkhb" event={"ID":"727c4472-c2ef-41fb-a337-a9ec52aba849","Type":"ContainerStarted","Data":"cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799"} Dec 05 09:42:37 crc kubenswrapper[4815]: I1205 09:42:37.145701 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9mkhb" podStartSLOduration=2.7343651639999997 podStartE2EDuration="5.145680063s" podCreationTimestamp="2025-12-05 09:42:32 +0000 UTC" firstStartedPulling="2025-12-05 09:42:34.088018191 +0000 UTC m=+2152.966625038" lastFinishedPulling="2025-12-05 09:42:36.49933309 +0000 UTC m=+2155.377939937" observedRunningTime="2025-12-05 09:42:37.137817028 +0000 UTC m=+2156.016423865" watchObservedRunningTime="2025-12-05 09:42:37.145680063 +0000 UTC m=+2156.024286900" Dec 05 09:42:39 crc kubenswrapper[4815]: I1205 09:42:39.140550 4815 generic.go:334] "Generic (PLEG): container finished" podID="962a4d01-45df-47a9-a9da-b36db0314c2d" containerID="e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397" exitCode=0 Dec 05 09:42:39 crc kubenswrapper[4815]: I1205 09:42:39.140654 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b55qd" event={"ID":"962a4d01-45df-47a9-a9da-b36db0314c2d","Type":"ContainerDied","Data":"e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397"} Dec 05 09:42:40 crc kubenswrapper[4815]: I1205 09:42:40.151825 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b55qd" event={"ID":"962a4d01-45df-47a9-a9da-b36db0314c2d","Type":"ContainerStarted","Data":"abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7"} Dec 05 09:42:40 crc kubenswrapper[4815]: I1205 09:42:40.181211 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-b55qd" podStartSLOduration=2.753818271 podStartE2EDuration="6.181193372s" podCreationTimestamp="2025-12-05 09:42:34 +0000 UTC" firstStartedPulling="2025-12-05 09:42:36.105498875 +0000 UTC m=+2154.984105712" lastFinishedPulling="2025-12-05 09:42:39.532873976 +0000 UTC m=+2158.411480813" observedRunningTime="2025-12-05 09:42:40.179234939 +0000 UTC m=+2159.057841776" watchObservedRunningTime="2025-12-05 09:42:40.181193372 +0000 UTC m=+2159.059800209" Dec 05 09:42:42 crc kubenswrapper[4815]: I1205 09:42:42.562798 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:42 crc kubenswrapper[4815]: I1205 09:42:42.564077 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:42 crc kubenswrapper[4815]: I1205 09:42:42.611004 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:43 crc kubenswrapper[4815]: I1205 09:42:43.261286 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:44 crc kubenswrapper[4815]: I1205 09:42:44.195073 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mkhb"] Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.144098 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.144551 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.200613 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.220534 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9mkhb" podUID="727c4472-c2ef-41fb-a337-a9ec52aba849" containerName="registry-server" containerID="cri-o://cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799" gracePeriod=2 Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.273752 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.744108 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.767002 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-utilities\") pod \"727c4472-c2ef-41fb-a337-a9ec52aba849\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.767045 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-catalog-content\") pod \"727c4472-c2ef-41fb-a337-a9ec52aba849\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.767084 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htcxc\" (UniqueName: \"kubernetes.io/projected/727c4472-c2ef-41fb-a337-a9ec52aba849-kube-api-access-htcxc\") pod \"727c4472-c2ef-41fb-a337-a9ec52aba849\" (UID: \"727c4472-c2ef-41fb-a337-a9ec52aba849\") " Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.770292 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-utilities" (OuterVolumeSpecName: "utilities") pod "727c4472-c2ef-41fb-a337-a9ec52aba849" (UID: "727c4472-c2ef-41fb-a337-a9ec52aba849"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.780285 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/727c4472-c2ef-41fb-a337-a9ec52aba849-kube-api-access-htcxc" (OuterVolumeSpecName: "kube-api-access-htcxc") pod "727c4472-c2ef-41fb-a337-a9ec52aba849" (UID: "727c4472-c2ef-41fb-a337-a9ec52aba849"). InnerVolumeSpecName "kube-api-access-htcxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.789916 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "727c4472-c2ef-41fb-a337-a9ec52aba849" (UID: "727c4472-c2ef-41fb-a337-a9ec52aba849"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.868677 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.868711 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/727c4472-c2ef-41fb-a337-a9ec52aba849-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:42:45 crc kubenswrapper[4815]: I1205 09:42:45.868722 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htcxc\" (UniqueName: \"kubernetes.io/projected/727c4472-c2ef-41fb-a337-a9ec52aba849-kube-api-access-htcxc\") on node \"crc\" DevicePath \"\"" Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.230611 4815 generic.go:334] "Generic (PLEG): container finished" podID="727c4472-c2ef-41fb-a337-a9ec52aba849" containerID="cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799" exitCode=0 Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.230663 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9mkhb" Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.230686 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mkhb" event={"ID":"727c4472-c2ef-41fb-a337-a9ec52aba849","Type":"ContainerDied","Data":"cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799"} Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.230716 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9mkhb" event={"ID":"727c4472-c2ef-41fb-a337-a9ec52aba849","Type":"ContainerDied","Data":"0157598ba1a4e8fd8393f9c8e62e758a228b410b3cbf864d9f75a5405f227971"} Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.230752 4815 scope.go:117] "RemoveContainer" containerID="cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799" Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.247444 4815 scope.go:117] "RemoveContainer" containerID="5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee" Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.279002 4815 scope.go:117] "RemoveContainer" containerID="b236d9ec888fa2714fc04a708cb6a35e20472f064dae0adf1d514f871d5f4e47" Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.293861 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mkhb"] Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.302828 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9mkhb"] Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.320814 4815 scope.go:117] "RemoveContainer" containerID="cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799" Dec 05 09:42:46 crc kubenswrapper[4815]: E1205 09:42:46.321389 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799\": container with ID starting with cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799 not found: ID does not exist" containerID="cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799" Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.321422 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799"} err="failed to get container status \"cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799\": rpc error: code = NotFound desc = could not find container \"cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799\": container with ID starting with cc842de9382ad9b175b811c1ecc73815d8f9efd0db41561c32f51c4f7ecac799 not found: ID does not exist" Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.321447 4815 scope.go:117] "RemoveContainer" containerID="5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee" Dec 05 09:42:46 crc kubenswrapper[4815]: E1205 09:42:46.322759 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee\": container with ID starting with 5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee not found: ID does not exist" containerID="5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee" Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.322781 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee"} err="failed to get container status \"5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee\": rpc error: code = NotFound desc = could not find container \"5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee\": container with ID starting with 5dc7926fadd5c661594f24c45a55a52898a099c97528dec50dbe8962acbb43ee not found: ID does not exist" Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.322797 4815 scope.go:117] "RemoveContainer" containerID="b236d9ec888fa2714fc04a708cb6a35e20472f064dae0adf1d514f871d5f4e47" Dec 05 09:42:46 crc kubenswrapper[4815]: E1205 09:42:46.323112 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b236d9ec888fa2714fc04a708cb6a35e20472f064dae0adf1d514f871d5f4e47\": container with ID starting with b236d9ec888fa2714fc04a708cb6a35e20472f064dae0adf1d514f871d5f4e47 not found: ID does not exist" containerID="b236d9ec888fa2714fc04a708cb6a35e20472f064dae0adf1d514f871d5f4e47" Dec 05 09:42:46 crc kubenswrapper[4815]: I1205 09:42:46.323165 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b236d9ec888fa2714fc04a708cb6a35e20472f064dae0adf1d514f871d5f4e47"} err="failed to get container status \"b236d9ec888fa2714fc04a708cb6a35e20472f064dae0adf1d514f871d5f4e47\": rpc error: code = NotFound desc = could not find container \"b236d9ec888fa2714fc04a708cb6a35e20472f064dae0adf1d514f871d5f4e47\": container with ID starting with b236d9ec888fa2714fc04a708cb6a35e20472f064dae0adf1d514f871d5f4e47 not found: ID does not exist" Dec 05 09:42:47 crc kubenswrapper[4815]: I1205 09:42:47.399876 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b55qd"] Dec 05 09:42:47 crc kubenswrapper[4815]: I1205 09:42:47.400382 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-b55qd" podUID="962a4d01-45df-47a9-a9da-b36db0314c2d" containerName="registry-server" containerID="cri-o://abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7" gracePeriod=2 Dec 05 09:42:47 crc kubenswrapper[4815]: I1205 09:42:47.432713 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="727c4472-c2ef-41fb-a337-a9ec52aba849" path="/var/lib/kubelet/pods/727c4472-c2ef-41fb-a337-a9ec52aba849/volumes" Dec 05 09:42:47 crc kubenswrapper[4815]: I1205 09:42:47.940114 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.018340 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-catalog-content\") pod \"962a4d01-45df-47a9-a9da-b36db0314c2d\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.018423 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6hr5\" (UniqueName: \"kubernetes.io/projected/962a4d01-45df-47a9-a9da-b36db0314c2d-kube-api-access-z6hr5\") pod \"962a4d01-45df-47a9-a9da-b36db0314c2d\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.018472 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-utilities\") pod \"962a4d01-45df-47a9-a9da-b36db0314c2d\" (UID: \"962a4d01-45df-47a9-a9da-b36db0314c2d\") " Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.019587 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-utilities" (OuterVolumeSpecName: "utilities") pod "962a4d01-45df-47a9-a9da-b36db0314c2d" (UID: "962a4d01-45df-47a9-a9da-b36db0314c2d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.026529 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/962a4d01-45df-47a9-a9da-b36db0314c2d-kube-api-access-z6hr5" (OuterVolumeSpecName: "kube-api-access-z6hr5") pod "962a4d01-45df-47a9-a9da-b36db0314c2d" (UID: "962a4d01-45df-47a9-a9da-b36db0314c2d"). InnerVolumeSpecName "kube-api-access-z6hr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.076129 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "962a4d01-45df-47a9-a9da-b36db0314c2d" (UID: "962a4d01-45df-47a9-a9da-b36db0314c2d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.126937 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.126989 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6hr5\" (UniqueName: \"kubernetes.io/projected/962a4d01-45df-47a9-a9da-b36db0314c2d-kube-api-access-z6hr5\") on node \"crc\" DevicePath \"\"" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.127002 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/962a4d01-45df-47a9-a9da-b36db0314c2d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.252339 4815 generic.go:334] "Generic (PLEG): container finished" podID="962a4d01-45df-47a9-a9da-b36db0314c2d" containerID="abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7" exitCode=0 Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.252393 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b55qd" event={"ID":"962a4d01-45df-47a9-a9da-b36db0314c2d","Type":"ContainerDied","Data":"abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7"} Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.252411 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-b55qd" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.252428 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-b55qd" event={"ID":"962a4d01-45df-47a9-a9da-b36db0314c2d","Type":"ContainerDied","Data":"4c3db9748ebecf1546d178476294c5fe2cfe8be3d56416ce36fb96e16bae6d9a"} Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.252449 4815 scope.go:117] "RemoveContainer" containerID="abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.274782 4815 scope.go:117] "RemoveContainer" containerID="e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.288474 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-b55qd"] Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.299184 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-b55qd"] Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.310007 4815 scope.go:117] "RemoveContainer" containerID="a40ca4d2c88f26dd2a7290cde852a124a7cf45063ee99e451f8da186df77ddee" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.355365 4815 scope.go:117] "RemoveContainer" containerID="abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7" Dec 05 09:42:48 crc kubenswrapper[4815]: E1205 09:42:48.356017 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7\": container with ID starting with abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7 not found: ID does not exist" containerID="abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.356063 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7"} err="failed to get container status \"abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7\": rpc error: code = NotFound desc = could not find container \"abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7\": container with ID starting with abbadbee6302359ccceae6e4ceb1ed6cd5b9c2ab19f4dbf15b265fe256a8baa7 not found: ID does not exist" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.356098 4815 scope.go:117] "RemoveContainer" containerID="e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397" Dec 05 09:42:48 crc kubenswrapper[4815]: E1205 09:42:48.356592 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397\": container with ID starting with e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397 not found: ID does not exist" containerID="e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.356631 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397"} err="failed to get container status \"e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397\": rpc error: code = NotFound desc = could not find container \"e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397\": container with ID starting with e7302b189e9fd4e1e0786e91886283a0ad745c6ef7e905f60f3a2bc802de5397 not found: ID does not exist" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.356661 4815 scope.go:117] "RemoveContainer" containerID="a40ca4d2c88f26dd2a7290cde852a124a7cf45063ee99e451f8da186df77ddee" Dec 05 09:42:48 crc kubenswrapper[4815]: E1205 09:42:48.357066 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a40ca4d2c88f26dd2a7290cde852a124a7cf45063ee99e451f8da186df77ddee\": container with ID starting with a40ca4d2c88f26dd2a7290cde852a124a7cf45063ee99e451f8da186df77ddee not found: ID does not exist" containerID="a40ca4d2c88f26dd2a7290cde852a124a7cf45063ee99e451f8da186df77ddee" Dec 05 09:42:48 crc kubenswrapper[4815]: I1205 09:42:48.357146 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a40ca4d2c88f26dd2a7290cde852a124a7cf45063ee99e451f8da186df77ddee"} err="failed to get container status \"a40ca4d2c88f26dd2a7290cde852a124a7cf45063ee99e451f8da186df77ddee\": rpc error: code = NotFound desc = could not find container \"a40ca4d2c88f26dd2a7290cde852a124a7cf45063ee99e451f8da186df77ddee\": container with ID starting with a40ca4d2c88f26dd2a7290cde852a124a7cf45063ee99e451f8da186df77ddee not found: ID does not exist" Dec 05 09:42:49 crc kubenswrapper[4815]: I1205 09:42:49.433649 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="962a4d01-45df-47a9-a9da-b36db0314c2d" path="/var/lib/kubelet/pods/962a4d01-45df-47a9-a9da-b36db0314c2d/volumes" Dec 05 09:44:20 crc kubenswrapper[4815]: I1205 09:44:20.192681 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:44:20 crc kubenswrapper[4815]: I1205 09:44:20.194357 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:44:50 crc kubenswrapper[4815]: I1205 09:44:50.191984 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:44:50 crc kubenswrapper[4815]: I1205 09:44:50.192541 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.339724 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5spt4"] Dec 05 09:44:53 crc kubenswrapper[4815]: E1205 09:44:53.342215 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727c4472-c2ef-41fb-a337-a9ec52aba849" containerName="registry-server" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.342336 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="727c4472-c2ef-41fb-a337-a9ec52aba849" containerName="registry-server" Dec 05 09:44:53 crc kubenswrapper[4815]: E1205 09:44:53.342434 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727c4472-c2ef-41fb-a337-a9ec52aba849" containerName="extract-content" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.342536 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="727c4472-c2ef-41fb-a337-a9ec52aba849" containerName="extract-content" Dec 05 09:44:53 crc kubenswrapper[4815]: E1205 09:44:53.342621 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="962a4d01-45df-47a9-a9da-b36db0314c2d" containerName="extract-utilities" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.342685 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="962a4d01-45df-47a9-a9da-b36db0314c2d" containerName="extract-utilities" Dec 05 09:44:53 crc kubenswrapper[4815]: E1205 09:44:53.342772 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="962a4d01-45df-47a9-a9da-b36db0314c2d" containerName="registry-server" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.342852 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="962a4d01-45df-47a9-a9da-b36db0314c2d" containerName="registry-server" Dec 05 09:44:53 crc kubenswrapper[4815]: E1205 09:44:53.342931 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="962a4d01-45df-47a9-a9da-b36db0314c2d" containerName="extract-content" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.343000 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="962a4d01-45df-47a9-a9da-b36db0314c2d" containerName="extract-content" Dec 05 09:44:53 crc kubenswrapper[4815]: E1205 09:44:53.343131 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="727c4472-c2ef-41fb-a337-a9ec52aba849" containerName="extract-utilities" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.343199 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="727c4472-c2ef-41fb-a337-a9ec52aba849" containerName="extract-utilities" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.345098 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="727c4472-c2ef-41fb-a337-a9ec52aba849" containerName="registry-server" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.345217 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="962a4d01-45df-47a9-a9da-b36db0314c2d" containerName="registry-server" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.347650 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.354685 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5spt4"] Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.407572 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glcwd\" (UniqueName: \"kubernetes.io/projected/47817848-221d-42cd-9feb-321d9dcdca05-kube-api-access-glcwd\") pod \"community-operators-5spt4\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.407656 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-catalog-content\") pod \"community-operators-5spt4\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.407697 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-utilities\") pod \"community-operators-5spt4\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.510040 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glcwd\" (UniqueName: \"kubernetes.io/projected/47817848-221d-42cd-9feb-321d9dcdca05-kube-api-access-glcwd\") pod \"community-operators-5spt4\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.510137 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-catalog-content\") pod \"community-operators-5spt4\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.510171 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-utilities\") pod \"community-operators-5spt4\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.513060 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-catalog-content\") pod \"community-operators-5spt4\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.513302 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-utilities\") pod \"community-operators-5spt4\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.538224 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glcwd\" (UniqueName: \"kubernetes.io/projected/47817848-221d-42cd-9feb-321d9dcdca05-kube-api-access-glcwd\") pod \"community-operators-5spt4\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:44:53 crc kubenswrapper[4815]: I1205 09:44:53.673318 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:44:54 crc kubenswrapper[4815]: I1205 09:44:54.331913 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5spt4"] Dec 05 09:44:54 crc kubenswrapper[4815]: I1205 09:44:54.435562 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5spt4" event={"ID":"47817848-221d-42cd-9feb-321d9dcdca05","Type":"ContainerStarted","Data":"bd8026b597f3e8bcc228a545db9fa74338467f715821d523fca6e43dd8da4c2f"} Dec 05 09:44:55 crc kubenswrapper[4815]: I1205 09:44:55.446664 4815 generic.go:334] "Generic (PLEG): container finished" podID="47817848-221d-42cd-9feb-321d9dcdca05" containerID="c245d15f072ee053b0aaedd21c247f360212236b5fc542d9abd73789123d8ab5" exitCode=0 Dec 05 09:44:55 crc kubenswrapper[4815]: I1205 09:44:55.446765 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5spt4" event={"ID":"47817848-221d-42cd-9feb-321d9dcdca05","Type":"ContainerDied","Data":"c245d15f072ee053b0aaedd21c247f360212236b5fc542d9abd73789123d8ab5"} Dec 05 09:44:55 crc kubenswrapper[4815]: I1205 09:44:55.448556 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:44:57 crc kubenswrapper[4815]: I1205 09:44:57.478399 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5spt4" event={"ID":"47817848-221d-42cd-9feb-321d9dcdca05","Type":"ContainerStarted","Data":"a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483"} Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.574012 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf"] Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.576391 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.579203 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.579433 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.604023 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf"] Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.662231 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-secret-volume\") pod \"collect-profiles-29415465-dcxmf\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.662325 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7lf9s\" (UniqueName: \"kubernetes.io/projected/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-kube-api-access-7lf9s\") pod \"collect-profiles-29415465-dcxmf\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.662655 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-config-volume\") pod \"collect-profiles-29415465-dcxmf\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.764911 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7lf9s\" (UniqueName: \"kubernetes.io/projected/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-kube-api-access-7lf9s\") pod \"collect-profiles-29415465-dcxmf\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.765009 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-config-volume\") pod \"collect-profiles-29415465-dcxmf\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.765086 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-secret-volume\") pod \"collect-profiles-29415465-dcxmf\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.765993 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-config-volume\") pod \"collect-profiles-29415465-dcxmf\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.772666 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-secret-volume\") pod \"collect-profiles-29415465-dcxmf\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.784981 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7lf9s\" (UniqueName: \"kubernetes.io/projected/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-kube-api-access-7lf9s\") pod \"collect-profiles-29415465-dcxmf\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:00 crc kubenswrapper[4815]: I1205 09:45:00.911873 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:01 crc kubenswrapper[4815]: I1205 09:45:01.344030 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf"] Dec 05 09:45:01 crc kubenswrapper[4815]: I1205 09:45:01.514123 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" event={"ID":"d134fdd6-8898-49a1-8b2e-78d65bafd6d7","Type":"ContainerStarted","Data":"549c9a75a264cba58f3d845bfb07018c908f63b20cece90edad4579aef04d299"} Dec 05 09:45:03 crc kubenswrapper[4815]: I1205 09:45:03.537145 4815 generic.go:334] "Generic (PLEG): container finished" podID="47817848-221d-42cd-9feb-321d9dcdca05" containerID="a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483" exitCode=0 Dec 05 09:45:03 crc kubenswrapper[4815]: I1205 09:45:03.537477 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5spt4" event={"ID":"47817848-221d-42cd-9feb-321d9dcdca05","Type":"ContainerDied","Data":"a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483"} Dec 05 09:45:06 crc kubenswrapper[4815]: I1205 09:45:06.566585 4815 generic.go:334] "Generic (PLEG): container finished" podID="d134fdd6-8898-49a1-8b2e-78d65bafd6d7" containerID="d934e352a34ce974070c6e3d67fc9faeca7b4495d40edce7819650ea82d899b9" exitCode=0 Dec 05 09:45:06 crc kubenswrapper[4815]: I1205 09:45:06.566639 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" event={"ID":"d134fdd6-8898-49a1-8b2e-78d65bafd6d7","Type":"ContainerDied","Data":"d934e352a34ce974070c6e3d67fc9faeca7b4495d40edce7819650ea82d899b9"} Dec 05 09:45:07 crc kubenswrapper[4815]: I1205 09:45:07.578794 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5spt4" event={"ID":"47817848-221d-42cd-9feb-321d9dcdca05","Type":"ContainerStarted","Data":"e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0"} Dec 05 09:45:07 crc kubenswrapper[4815]: I1205 09:45:07.608741 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5spt4" podStartSLOduration=3.650061713 podStartE2EDuration="14.608673331s" podCreationTimestamp="2025-12-05 09:44:53 +0000 UTC" firstStartedPulling="2025-12-05 09:44:55.448218462 +0000 UTC m=+2294.326825299" lastFinishedPulling="2025-12-05 09:45:06.40683008 +0000 UTC m=+2305.285436917" observedRunningTime="2025-12-05 09:45:07.604348164 +0000 UTC m=+2306.482955011" watchObservedRunningTime="2025-12-05 09:45:07.608673331 +0000 UTC m=+2306.487280168" Dec 05 09:45:07 crc kubenswrapper[4815]: I1205 09:45:07.935585 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.006728 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-secret-volume\") pod \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.006993 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7lf9s\" (UniqueName: \"kubernetes.io/projected/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-kube-api-access-7lf9s\") pod \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.007156 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-config-volume\") pod \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\" (UID: \"d134fdd6-8898-49a1-8b2e-78d65bafd6d7\") " Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.007931 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-config-volume" (OuterVolumeSpecName: "config-volume") pod "d134fdd6-8898-49a1-8b2e-78d65bafd6d7" (UID: "d134fdd6-8898-49a1-8b2e-78d65bafd6d7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.013877 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-kube-api-access-7lf9s" (OuterVolumeSpecName: "kube-api-access-7lf9s") pod "d134fdd6-8898-49a1-8b2e-78d65bafd6d7" (UID: "d134fdd6-8898-49a1-8b2e-78d65bafd6d7"). InnerVolumeSpecName "kube-api-access-7lf9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.014002 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d134fdd6-8898-49a1-8b2e-78d65bafd6d7" (UID: "d134fdd6-8898-49a1-8b2e-78d65bafd6d7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.109766 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.109803 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7lf9s\" (UniqueName: \"kubernetes.io/projected/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-kube-api-access-7lf9s\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.109813 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d134fdd6-8898-49a1-8b2e-78d65bafd6d7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.587605 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" event={"ID":"d134fdd6-8898-49a1-8b2e-78d65bafd6d7","Type":"ContainerDied","Data":"549c9a75a264cba58f3d845bfb07018c908f63b20cece90edad4579aef04d299"} Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.587644 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="549c9a75a264cba58f3d845bfb07018c908f63b20cece90edad4579aef04d299" Dec 05 09:45:08 crc kubenswrapper[4815]: I1205 09:45:08.587662 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf" Dec 05 09:45:09 crc kubenswrapper[4815]: I1205 09:45:09.023357 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w"] Dec 05 09:45:09 crc kubenswrapper[4815]: I1205 09:45:09.031032 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415420-98n6w"] Dec 05 09:45:09 crc kubenswrapper[4815]: I1205 09:45:09.447466 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b78558ed-6530-417a-a88e-a17d601593b0" path="/var/lib/kubelet/pods/b78558ed-6530-417a-a88e-a17d601593b0/volumes" Dec 05 09:45:13 crc kubenswrapper[4815]: I1205 09:45:13.674366 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:45:13 crc kubenswrapper[4815]: I1205 09:45:13.674783 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:45:13 crc kubenswrapper[4815]: I1205 09:45:13.724227 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:45:14 crc kubenswrapper[4815]: I1205 09:45:14.683688 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:45:14 crc kubenswrapper[4815]: I1205 09:45:14.735757 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5spt4"] Dec 05 09:45:16 crc kubenswrapper[4815]: I1205 09:45:16.653957 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5spt4" podUID="47817848-221d-42cd-9feb-321d9dcdca05" containerName="registry-server" containerID="cri-o://e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0" gracePeriod=2 Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.585632 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.701780 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-utilities\") pod \"47817848-221d-42cd-9feb-321d9dcdca05\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.702038 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-glcwd\" (UniqueName: \"kubernetes.io/projected/47817848-221d-42cd-9feb-321d9dcdca05-kube-api-access-glcwd\") pod \"47817848-221d-42cd-9feb-321d9dcdca05\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.702061 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-catalog-content\") pod \"47817848-221d-42cd-9feb-321d9dcdca05\" (UID: \"47817848-221d-42cd-9feb-321d9dcdca05\") " Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.703063 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-utilities" (OuterVolumeSpecName: "utilities") pod "47817848-221d-42cd-9feb-321d9dcdca05" (UID: "47817848-221d-42cd-9feb-321d9dcdca05"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.708348 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47817848-221d-42cd-9feb-321d9dcdca05-kube-api-access-glcwd" (OuterVolumeSpecName: "kube-api-access-glcwd") pod "47817848-221d-42cd-9feb-321d9dcdca05" (UID: "47817848-221d-42cd-9feb-321d9dcdca05"). InnerVolumeSpecName "kube-api-access-glcwd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.715353 4815 generic.go:334] "Generic (PLEG): container finished" podID="47817848-221d-42cd-9feb-321d9dcdca05" containerID="e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0" exitCode=0 Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.715390 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5spt4" event={"ID":"47817848-221d-42cd-9feb-321d9dcdca05","Type":"ContainerDied","Data":"e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0"} Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.715416 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5spt4" event={"ID":"47817848-221d-42cd-9feb-321d9dcdca05","Type":"ContainerDied","Data":"bd8026b597f3e8bcc228a545db9fa74338467f715821d523fca6e43dd8da4c2f"} Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.715456 4815 scope.go:117] "RemoveContainer" containerID="e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.715635 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5spt4" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.764943 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47817848-221d-42cd-9feb-321d9dcdca05" (UID: "47817848-221d-42cd-9feb-321d9dcdca05"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.765137 4815 scope.go:117] "RemoveContainer" containerID="a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.783762 4815 scope.go:117] "RemoveContainer" containerID="c245d15f072ee053b0aaedd21c247f360212236b5fc542d9abd73789123d8ab5" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.803566 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.803620 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-glcwd\" (UniqueName: \"kubernetes.io/projected/47817848-221d-42cd-9feb-321d9dcdca05-kube-api-access-glcwd\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.803784 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47817848-221d-42cd-9feb-321d9dcdca05-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.819236 4815 scope.go:117] "RemoveContainer" containerID="e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0" Dec 05 09:45:17 crc kubenswrapper[4815]: E1205 09:45:17.820009 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0\": container with ID starting with e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0 not found: ID does not exist" containerID="e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.820100 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0"} err="failed to get container status \"e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0\": rpc error: code = NotFound desc = could not find container \"e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0\": container with ID starting with e69735230eb162eac655eeec465f11989006c4a255ff30ea34265d16f301e2b0 not found: ID does not exist" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.820131 4815 scope.go:117] "RemoveContainer" containerID="a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483" Dec 05 09:45:17 crc kubenswrapper[4815]: E1205 09:45:17.820767 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483\": container with ID starting with a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483 not found: ID does not exist" containerID="a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.820800 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483"} err="failed to get container status \"a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483\": rpc error: code = NotFound desc = could not find container \"a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483\": container with ID starting with a6945fed2ecdabdf88b3644aa9850e14e6186b30023093de5b89576ab8345483 not found: ID does not exist" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.820837 4815 scope.go:117] "RemoveContainer" containerID="c245d15f072ee053b0aaedd21c247f360212236b5fc542d9abd73789123d8ab5" Dec 05 09:45:17 crc kubenswrapper[4815]: E1205 09:45:17.821132 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c245d15f072ee053b0aaedd21c247f360212236b5fc542d9abd73789123d8ab5\": container with ID starting with c245d15f072ee053b0aaedd21c247f360212236b5fc542d9abd73789123d8ab5 not found: ID does not exist" containerID="c245d15f072ee053b0aaedd21c247f360212236b5fc542d9abd73789123d8ab5" Dec 05 09:45:17 crc kubenswrapper[4815]: I1205 09:45:17.821157 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c245d15f072ee053b0aaedd21c247f360212236b5fc542d9abd73789123d8ab5"} err="failed to get container status \"c245d15f072ee053b0aaedd21c247f360212236b5fc542d9abd73789123d8ab5\": rpc error: code = NotFound desc = could not find container \"c245d15f072ee053b0aaedd21c247f360212236b5fc542d9abd73789123d8ab5\": container with ID starting with c245d15f072ee053b0aaedd21c247f360212236b5fc542d9abd73789123d8ab5 not found: ID does not exist" Dec 05 09:45:18 crc kubenswrapper[4815]: I1205 09:45:18.054931 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5spt4"] Dec 05 09:45:18 crc kubenswrapper[4815]: I1205 09:45:18.062316 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5spt4"] Dec 05 09:45:19 crc kubenswrapper[4815]: I1205 09:45:19.432312 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47817848-221d-42cd-9feb-321d9dcdca05" path="/var/lib/kubelet/pods/47817848-221d-42cd-9feb-321d9dcdca05/volumes" Dec 05 09:45:20 crc kubenswrapper[4815]: I1205 09:45:20.192199 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:45:20 crc kubenswrapper[4815]: I1205 09:45:20.192556 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:45:20 crc kubenswrapper[4815]: I1205 09:45:20.192791 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:45:20 crc kubenswrapper[4815]: I1205 09:45:20.193928 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:45:20 crc kubenswrapper[4815]: I1205 09:45:20.194132 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" gracePeriod=600 Dec 05 09:45:20 crc kubenswrapper[4815]: E1205 09:45:20.327005 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:45:20 crc kubenswrapper[4815]: I1205 09:45:20.762453 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" exitCode=0 Dec 05 09:45:20 crc kubenswrapper[4815]: I1205 09:45:20.762521 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372"} Dec 05 09:45:20 crc kubenswrapper[4815]: I1205 09:45:20.762583 4815 scope.go:117] "RemoveContainer" containerID="1db619e28f79e22d68fc8699b12d64acf5e1c448e650971b8846c7b1098b9c14" Dec 05 09:45:20 crc kubenswrapper[4815]: I1205 09:45:20.763185 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:45:20 crc kubenswrapper[4815]: E1205 09:45:20.763781 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:45:34 crc kubenswrapper[4815]: I1205 09:45:34.419712 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:45:34 crc kubenswrapper[4815]: E1205 09:45:34.421020 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.629331 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-ttt5l"] Dec 05 09:45:47 crc kubenswrapper[4815]: E1205 09:45:47.630314 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47817848-221d-42cd-9feb-321d9dcdca05" containerName="extract-content" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.630332 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="47817848-221d-42cd-9feb-321d9dcdca05" containerName="extract-content" Dec 05 09:45:47 crc kubenswrapper[4815]: E1205 09:45:47.630349 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47817848-221d-42cd-9feb-321d9dcdca05" containerName="extract-utilities" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.630357 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="47817848-221d-42cd-9feb-321d9dcdca05" containerName="extract-utilities" Dec 05 09:45:47 crc kubenswrapper[4815]: E1205 09:45:47.630374 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d134fdd6-8898-49a1-8b2e-78d65bafd6d7" containerName="collect-profiles" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.630380 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d134fdd6-8898-49a1-8b2e-78d65bafd6d7" containerName="collect-profiles" Dec 05 09:45:47 crc kubenswrapper[4815]: E1205 09:45:47.630395 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47817848-221d-42cd-9feb-321d9dcdca05" containerName="registry-server" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.630401 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="47817848-221d-42cd-9feb-321d9dcdca05" containerName="registry-server" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.630625 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="47817848-221d-42cd-9feb-321d9dcdca05" containerName="registry-server" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.630652 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d134fdd6-8898-49a1-8b2e-78d65bafd6d7" containerName="collect-profiles" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.631940 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.646038 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ttt5l"] Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.737708 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwk5j\" (UniqueName: \"kubernetes.io/projected/422fe87d-262f-408f-b59c-d7071acbd202-kube-api-access-jwk5j\") pod \"redhat-operators-ttt5l\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.737786 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-catalog-content\") pod \"redhat-operators-ttt5l\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.737857 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-utilities\") pod \"redhat-operators-ttt5l\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.839543 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwk5j\" (UniqueName: \"kubernetes.io/projected/422fe87d-262f-408f-b59c-d7071acbd202-kube-api-access-jwk5j\") pod \"redhat-operators-ttt5l\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.839606 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-catalog-content\") pod \"redhat-operators-ttt5l\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.839645 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-utilities\") pod \"redhat-operators-ttt5l\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.840201 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-catalog-content\") pod \"redhat-operators-ttt5l\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.840216 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-utilities\") pod \"redhat-operators-ttt5l\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.874701 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwk5j\" (UniqueName: \"kubernetes.io/projected/422fe87d-262f-408f-b59c-d7071acbd202-kube-api-access-jwk5j\") pod \"redhat-operators-ttt5l\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:45:47 crc kubenswrapper[4815]: I1205 09:45:47.952321 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:45:48 crc kubenswrapper[4815]: I1205 09:45:48.419021 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:45:48 crc kubenswrapper[4815]: E1205 09:45:48.419915 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:45:48 crc kubenswrapper[4815]: I1205 09:45:48.519560 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-ttt5l"] Dec 05 09:45:49 crc kubenswrapper[4815]: I1205 09:45:49.238869 4815 generic.go:334] "Generic (PLEG): container finished" podID="422fe87d-262f-408f-b59c-d7071acbd202" containerID="111c93c5bae54b81e6814af92b976f884ed77b13f250044656826cadabcc18f5" exitCode=0 Dec 05 09:45:49 crc kubenswrapper[4815]: I1205 09:45:49.239300 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ttt5l" event={"ID":"422fe87d-262f-408f-b59c-d7071acbd202","Type":"ContainerDied","Data":"111c93c5bae54b81e6814af92b976f884ed77b13f250044656826cadabcc18f5"} Dec 05 09:45:49 crc kubenswrapper[4815]: I1205 09:45:49.239329 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ttt5l" event={"ID":"422fe87d-262f-408f-b59c-d7071acbd202","Type":"ContainerStarted","Data":"41a1989eda291062a43a81e9488475d35a814e50371ccc946e8783b9077008a6"} Dec 05 09:45:51 crc kubenswrapper[4815]: I1205 09:45:51.263449 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ttt5l" event={"ID":"422fe87d-262f-408f-b59c-d7071acbd202","Type":"ContainerStarted","Data":"4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319"} Dec 05 09:45:57 crc kubenswrapper[4815]: I1205 09:45:57.314926 4815 generic.go:334] "Generic (PLEG): container finished" podID="422fe87d-262f-408f-b59c-d7071acbd202" containerID="4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319" exitCode=0 Dec 05 09:45:57 crc kubenswrapper[4815]: I1205 09:45:57.315406 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ttt5l" event={"ID":"422fe87d-262f-408f-b59c-d7071acbd202","Type":"ContainerDied","Data":"4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319"} Dec 05 09:45:59 crc kubenswrapper[4815]: I1205 09:45:59.320468 4815 scope.go:117] "RemoveContainer" containerID="e48039f48a5987476a6d576e5e98e3bc774c6a2baa8b0ad29636a6cccc757806" Dec 05 09:45:59 crc kubenswrapper[4815]: I1205 09:45:59.344742 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ttt5l" event={"ID":"422fe87d-262f-408f-b59c-d7071acbd202","Type":"ContainerStarted","Data":"8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e"} Dec 05 09:45:59 crc kubenswrapper[4815]: I1205 09:45:59.368700 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-ttt5l" podStartSLOduration=2.544377054 podStartE2EDuration="12.368682651s" podCreationTimestamp="2025-12-05 09:45:47 +0000 UTC" firstStartedPulling="2025-12-05 09:45:49.244689145 +0000 UTC m=+2348.123295982" lastFinishedPulling="2025-12-05 09:45:59.068994742 +0000 UTC m=+2357.947601579" observedRunningTime="2025-12-05 09:45:59.364234279 +0000 UTC m=+2358.242841116" watchObservedRunningTime="2025-12-05 09:45:59.368682651 +0000 UTC m=+2358.247289488" Dec 05 09:46:01 crc kubenswrapper[4815]: I1205 09:46:01.424059 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:46:01 crc kubenswrapper[4815]: E1205 09:46:01.425564 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:46:07 crc kubenswrapper[4815]: I1205 09:46:07.952802 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:46:07 crc kubenswrapper[4815]: I1205 09:46:07.953301 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:46:07 crc kubenswrapper[4815]: I1205 09:46:07.995933 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:46:08 crc kubenswrapper[4815]: I1205 09:46:08.464066 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:46:08 crc kubenswrapper[4815]: I1205 09:46:08.510247 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ttt5l"] Dec 05 09:46:10 crc kubenswrapper[4815]: I1205 09:46:10.436232 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-ttt5l" podUID="422fe87d-262f-408f-b59c-d7071acbd202" containerName="registry-server" containerID="cri-o://8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e" gracePeriod=2 Dec 05 09:46:10 crc kubenswrapper[4815]: I1205 09:46:10.896354 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.038969 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-utilities\") pod \"422fe87d-262f-408f-b59c-d7071acbd202\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.039101 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jwk5j\" (UniqueName: \"kubernetes.io/projected/422fe87d-262f-408f-b59c-d7071acbd202-kube-api-access-jwk5j\") pod \"422fe87d-262f-408f-b59c-d7071acbd202\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.039175 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-catalog-content\") pod \"422fe87d-262f-408f-b59c-d7071acbd202\" (UID: \"422fe87d-262f-408f-b59c-d7071acbd202\") " Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.040037 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-utilities" (OuterVolumeSpecName: "utilities") pod "422fe87d-262f-408f-b59c-d7071acbd202" (UID: "422fe87d-262f-408f-b59c-d7071acbd202"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.044566 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/422fe87d-262f-408f-b59c-d7071acbd202-kube-api-access-jwk5j" (OuterVolumeSpecName: "kube-api-access-jwk5j") pod "422fe87d-262f-408f-b59c-d7071acbd202" (UID: "422fe87d-262f-408f-b59c-d7071acbd202"). InnerVolumeSpecName "kube-api-access-jwk5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.140920 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.140994 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jwk5j\" (UniqueName: \"kubernetes.io/projected/422fe87d-262f-408f-b59c-d7071acbd202-kube-api-access-jwk5j\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.158007 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "422fe87d-262f-408f-b59c-d7071acbd202" (UID: "422fe87d-262f-408f-b59c-d7071acbd202"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.242479 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/422fe87d-262f-408f-b59c-d7071acbd202-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.448114 4815 generic.go:334] "Generic (PLEG): container finished" podID="422fe87d-262f-408f-b59c-d7071acbd202" containerID="8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e" exitCode=0 Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.448165 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ttt5l" event={"ID":"422fe87d-262f-408f-b59c-d7071acbd202","Type":"ContainerDied","Data":"8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e"} Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.448200 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-ttt5l" event={"ID":"422fe87d-262f-408f-b59c-d7071acbd202","Type":"ContainerDied","Data":"41a1989eda291062a43a81e9488475d35a814e50371ccc946e8783b9077008a6"} Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.448221 4815 scope.go:117] "RemoveContainer" containerID="8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.448367 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-ttt5l" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.473748 4815 scope.go:117] "RemoveContainer" containerID="4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.486780 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-ttt5l"] Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.495183 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-ttt5l"] Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.513124 4815 scope.go:117] "RemoveContainer" containerID="111c93c5bae54b81e6814af92b976f884ed77b13f250044656826cadabcc18f5" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.540787 4815 scope.go:117] "RemoveContainer" containerID="8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e" Dec 05 09:46:11 crc kubenswrapper[4815]: E1205 09:46:11.541626 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e\": container with ID starting with 8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e not found: ID does not exist" containerID="8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.541660 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e"} err="failed to get container status \"8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e\": rpc error: code = NotFound desc = could not find container \"8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e\": container with ID starting with 8eccee354ffaf01bdb66c875d75677757decf09bd4d6677633d60bb0fd0e5b5e not found: ID does not exist" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.541682 4815 scope.go:117] "RemoveContainer" containerID="4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319" Dec 05 09:46:11 crc kubenswrapper[4815]: E1205 09:46:11.542035 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319\": container with ID starting with 4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319 not found: ID does not exist" containerID="4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.542086 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319"} err="failed to get container status \"4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319\": rpc error: code = NotFound desc = could not find container \"4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319\": container with ID starting with 4428f405f4bbe9db9e2a07d9d4810ebe2675c5ba5353f37688c25bc502f60319 not found: ID does not exist" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.542121 4815 scope.go:117] "RemoveContainer" containerID="111c93c5bae54b81e6814af92b976f884ed77b13f250044656826cadabcc18f5" Dec 05 09:46:11 crc kubenswrapper[4815]: E1205 09:46:11.542592 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"111c93c5bae54b81e6814af92b976f884ed77b13f250044656826cadabcc18f5\": container with ID starting with 111c93c5bae54b81e6814af92b976f884ed77b13f250044656826cadabcc18f5 not found: ID does not exist" containerID="111c93c5bae54b81e6814af92b976f884ed77b13f250044656826cadabcc18f5" Dec 05 09:46:11 crc kubenswrapper[4815]: I1205 09:46:11.542614 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"111c93c5bae54b81e6814af92b976f884ed77b13f250044656826cadabcc18f5"} err="failed to get container status \"111c93c5bae54b81e6814af92b976f884ed77b13f250044656826cadabcc18f5\": rpc error: code = NotFound desc = could not find container \"111c93c5bae54b81e6814af92b976f884ed77b13f250044656826cadabcc18f5\": container with ID starting with 111c93c5bae54b81e6814af92b976f884ed77b13f250044656826cadabcc18f5 not found: ID does not exist" Dec 05 09:46:13 crc kubenswrapper[4815]: I1205 09:46:13.418544 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:46:13 crc kubenswrapper[4815]: E1205 09:46:13.419960 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:46:13 crc kubenswrapper[4815]: I1205 09:46:13.429834 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="422fe87d-262f-408f-b59c-d7071acbd202" path="/var/lib/kubelet/pods/422fe87d-262f-408f-b59c-d7071acbd202/volumes" Dec 05 09:46:27 crc kubenswrapper[4815]: I1205 09:46:27.418678 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:46:27 crc kubenswrapper[4815]: E1205 09:46:27.419425 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:46:40 crc kubenswrapper[4815]: I1205 09:46:40.418903 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:46:40 crc kubenswrapper[4815]: E1205 09:46:40.419665 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:46:52 crc kubenswrapper[4815]: I1205 09:46:52.418567 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:46:52 crc kubenswrapper[4815]: E1205 09:46:52.419749 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:47:03 crc kubenswrapper[4815]: I1205 09:47:03.419740 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:47:03 crc kubenswrapper[4815]: E1205 09:47:03.423144 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:47:15 crc kubenswrapper[4815]: I1205 09:47:15.418775 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:47:15 crc kubenswrapper[4815]: E1205 09:47:15.419593 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:47:27 crc kubenswrapper[4815]: I1205 09:47:27.419905 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:47:27 crc kubenswrapper[4815]: E1205 09:47:27.420640 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:47:41 crc kubenswrapper[4815]: I1205 09:47:41.449466 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:47:41 crc kubenswrapper[4815]: E1205 09:47:41.452161 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:47:54 crc kubenswrapper[4815]: I1205 09:47:54.419053 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:47:54 crc kubenswrapper[4815]: E1205 09:47:54.419820 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:48:08 crc kubenswrapper[4815]: I1205 09:48:08.419009 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:48:08 crc kubenswrapper[4815]: E1205 09:48:08.419889 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:48:22 crc kubenswrapper[4815]: I1205 09:48:22.418611 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:48:22 crc kubenswrapper[4815]: E1205 09:48:22.419378 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:48:35 crc kubenswrapper[4815]: I1205 09:48:35.419324 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:48:35 crc kubenswrapper[4815]: E1205 09:48:35.420282 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:48:49 crc kubenswrapper[4815]: I1205 09:48:49.418841 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:48:49 crc kubenswrapper[4815]: E1205 09:48:49.419697 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:49:03 crc kubenswrapper[4815]: I1205 09:49:03.420401 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:49:03 crc kubenswrapper[4815]: E1205 09:49:03.421819 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:49:15 crc kubenswrapper[4815]: I1205 09:49:15.419370 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:49:15 crc kubenswrapper[4815]: E1205 09:49:15.420170 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:49:28 crc kubenswrapper[4815]: I1205 09:49:28.418722 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:49:28 crc kubenswrapper[4815]: E1205 09:49:28.419668 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.237073 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.245442 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.260524 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.269370 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.276602 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.283476 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.289542 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qcxk9"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.295763 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-h9m7k"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.302275 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.309309 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5wqfb"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.315981 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-j9pq6"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.323514 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-t7d9z"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.329862 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-cdwsd"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.335429 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.341168 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-925t6"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.346909 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-qthts"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.352814 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.359672 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-w4c7p"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.365781 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-hslnj"] Dec 05 09:49:34 crc kubenswrapper[4815]: I1205 09:49:34.371284 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-t7d9z"] Dec 05 09:49:35 crc kubenswrapper[4815]: I1205 09:49:35.428611 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2212e6a0-bdfa-41e1-8ff3-2b386ff23864" path="/var/lib/kubelet/pods/2212e6a0-bdfa-41e1-8ff3-2b386ff23864/volumes" Dec 05 09:49:35 crc kubenswrapper[4815]: I1205 09:49:35.429324 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a49a6d7-0038-47ab-acea-549233b7d78f" path="/var/lib/kubelet/pods/6a49a6d7-0038-47ab-acea-549233b7d78f/volumes" Dec 05 09:49:35 crc kubenswrapper[4815]: I1205 09:49:35.430006 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c167339-94cc-4c3c-ae9c-0e9b0ceb256b" path="/var/lib/kubelet/pods/7c167339-94cc-4c3c-ae9c-0e9b0ceb256b/volumes" Dec 05 09:49:35 crc kubenswrapper[4815]: I1205 09:49:35.430758 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83d6ef02-36e2-49bc-a1cd-5225686a0fdd" path="/var/lib/kubelet/pods/83d6ef02-36e2-49bc-a1cd-5225686a0fdd/volumes" Dec 05 09:49:35 crc kubenswrapper[4815]: I1205 09:49:35.431957 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b97a436-14ba-4b10-91b9-d7d90b73c62d" path="/var/lib/kubelet/pods/8b97a436-14ba-4b10-91b9-d7d90b73c62d/volumes" Dec 05 09:49:35 crc kubenswrapper[4815]: I1205 09:49:35.432621 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8fd95c26-42fe-4d0c-82a0-eca943ceaf79" path="/var/lib/kubelet/pods/8fd95c26-42fe-4d0c-82a0-eca943ceaf79/volumes" Dec 05 09:49:35 crc kubenswrapper[4815]: I1205 09:49:35.433267 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf14d613-f583-4916-9a10-07fe2de25e48" path="/var/lib/kubelet/pods/bf14d613-f583-4916-9a10-07fe2de25e48/volumes" Dec 05 09:49:35 crc kubenswrapper[4815]: I1205 09:49:35.434440 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c33d38f4-80e3-421d-ad3c-19fd6d7005f8" path="/var/lib/kubelet/pods/c33d38f4-80e3-421d-ad3c-19fd6d7005f8/volumes" Dec 05 09:49:35 crc kubenswrapper[4815]: I1205 09:49:35.435173 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c84f99f1-df65-4728-a42b-6b9d5658a228" path="/var/lib/kubelet/pods/c84f99f1-df65-4728-a42b-6b9d5658a228/volumes" Dec 05 09:49:35 crc kubenswrapper[4815]: I1205 09:49:35.435903 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f29fa7d7-8396-42f6-b244-ad560a622193" path="/var/lib/kubelet/pods/f29fa7d7-8396-42f6-b244-ad560a622193/volumes" Dec 05 09:49:39 crc kubenswrapper[4815]: I1205 09:49:39.418690 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:49:39 crc kubenswrapper[4815]: E1205 09:49:39.419591 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.111706 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s"] Dec 05 09:49:48 crc kubenswrapper[4815]: E1205 09:49:48.112792 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="422fe87d-262f-408f-b59c-d7071acbd202" containerName="registry-server" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.112819 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="422fe87d-262f-408f-b59c-d7071acbd202" containerName="registry-server" Dec 05 09:49:48 crc kubenswrapper[4815]: E1205 09:49:48.112909 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="422fe87d-262f-408f-b59c-d7071acbd202" containerName="extract-content" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.112919 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="422fe87d-262f-408f-b59c-d7071acbd202" containerName="extract-content" Dec 05 09:49:48 crc kubenswrapper[4815]: E1205 09:49:48.112935 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="422fe87d-262f-408f-b59c-d7071acbd202" containerName="extract-utilities" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.112945 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="422fe87d-262f-408f-b59c-d7071acbd202" containerName="extract-utilities" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.114360 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="422fe87d-262f-408f-b59c-d7071acbd202" containerName="registry-server" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.115293 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.117627 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.119580 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.119613 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.119733 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.119877 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.135266 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s"] Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.157773 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6n6h\" (UniqueName: \"kubernetes.io/projected/0e0848e7-a6e0-4672-bb56-2f40193be881-kube-api-access-r6n6h\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.157830 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.157856 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.157904 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.157940 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.259540 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.259594 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.259669 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.259735 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.259838 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6n6h\" (UniqueName: \"kubernetes.io/projected/0e0848e7-a6e0-4672-bb56-2f40193be881-kube-api-access-r6n6h\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.268219 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.268258 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.268326 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.272070 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.282096 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6n6h\" (UniqueName: \"kubernetes.io/projected/0e0848e7-a6e0-4672-bb56-2f40193be881-kube-api-access-r6n6h\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:48 crc kubenswrapper[4815]: I1205 09:49:48.439461 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:49:49 crc kubenswrapper[4815]: I1205 09:49:49.055549 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s"] Dec 05 09:49:49 crc kubenswrapper[4815]: I1205 09:49:49.402405 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" event={"ID":"0e0848e7-a6e0-4672-bb56-2f40193be881","Type":"ContainerStarted","Data":"70ad604d060f366fb10eaff6e335638595d54f13a688bf134b5dd01dda46bb1e"} Dec 05 09:49:50 crc kubenswrapper[4815]: I1205 09:49:50.412244 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" event={"ID":"0e0848e7-a6e0-4672-bb56-2f40193be881","Type":"ContainerStarted","Data":"99f8bed0a15e70d80321af9025f7de159997657d37007b8c09b6f1a65bbc85e7"} Dec 05 09:49:50 crc kubenswrapper[4815]: I1205 09:49:50.442027 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" podStartSLOduration=2.026992452 podStartE2EDuration="2.441992526s" podCreationTimestamp="2025-12-05 09:49:48 +0000 UTC" firstStartedPulling="2025-12-05 09:49:49.066780729 +0000 UTC m=+2587.945387566" lastFinishedPulling="2025-12-05 09:49:49.481780803 +0000 UTC m=+2588.360387640" observedRunningTime="2025-12-05 09:49:50.429517034 +0000 UTC m=+2589.308123861" watchObservedRunningTime="2025-12-05 09:49:50.441992526 +0000 UTC m=+2589.320599373" Dec 05 09:49:54 crc kubenswrapper[4815]: I1205 09:49:54.418455 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:49:54 crc kubenswrapper[4815]: E1205 09:49:54.419228 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:49:59 crc kubenswrapper[4815]: I1205 09:49:59.500962 4815 scope.go:117] "RemoveContainer" containerID="ea298044d76fcd6d85f472282a73221c7b63dbbc05cb96ad92241cf7cbf54eee" Dec 05 09:49:59 crc kubenswrapper[4815]: I1205 09:49:59.603129 4815 scope.go:117] "RemoveContainer" containerID="ff883880d2cd98b119cf093dcfe70e10b93acbe38cca5a0f6b807a841aa2903b" Dec 05 09:49:59 crc kubenswrapper[4815]: I1205 09:49:59.683020 4815 scope.go:117] "RemoveContainer" containerID="8a9e01c8f8d2a31a2064be85e6cc1a27e715db83bea23b7197664bedc70f37a6" Dec 05 09:49:59 crc kubenswrapper[4815]: I1205 09:49:59.729508 4815 scope.go:117] "RemoveContainer" containerID="7af5b7a27fd38a7e4b8639ab8e6661d8f65c768ce29bc1dce549957fdbcfe4c7" Dec 05 09:49:59 crc kubenswrapper[4815]: I1205 09:49:59.759930 4815 scope.go:117] "RemoveContainer" containerID="bc43e3d716f87a687679453b74e9ba1f7e6bd1ded31eb1d7e20d25467cc6015c" Dec 05 09:49:59 crc kubenswrapper[4815]: I1205 09:49:59.875342 4815 scope.go:117] "RemoveContainer" containerID="a20a9e53498757533027b60b62550f6e5fe588334703059cc6248532d860f214" Dec 05 09:49:59 crc kubenswrapper[4815]: I1205 09:49:59.947454 4815 scope.go:117] "RemoveContainer" containerID="05862728ea7ea39070e64b9e68cd5c3ade5b4016914f2ad349ccb13c78a598bd" Dec 05 09:49:59 crc kubenswrapper[4815]: I1205 09:49:59.981593 4815 scope.go:117] "RemoveContainer" containerID="9402a2252e503ec8bb7473f51f6206e13d02721eb4b4c0f25114c18ba020b380" Dec 05 09:50:00 crc kubenswrapper[4815]: I1205 09:50:00.020810 4815 scope.go:117] "RemoveContainer" containerID="2566e67858c16d008dcca1b1cac59338167110b391114555b4637ead1be6f595" Dec 05 09:50:00 crc kubenswrapper[4815]: I1205 09:50:00.074509 4815 scope.go:117] "RemoveContainer" containerID="46443ce923ce64a901168a9fc589b424e164362a99a53922866bb1d7c815e1d5" Dec 05 09:50:02 crc kubenswrapper[4815]: I1205 09:50:02.532935 4815 generic.go:334] "Generic (PLEG): container finished" podID="0e0848e7-a6e0-4672-bb56-2f40193be881" containerID="99f8bed0a15e70d80321af9025f7de159997657d37007b8c09b6f1a65bbc85e7" exitCode=0 Dec 05 09:50:02 crc kubenswrapper[4815]: I1205 09:50:02.533121 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" event={"ID":"0e0848e7-a6e0-4672-bb56-2f40193be881","Type":"ContainerDied","Data":"99f8bed0a15e70d80321af9025f7de159997657d37007b8c09b6f1a65bbc85e7"} Dec 05 09:50:03 crc kubenswrapper[4815]: I1205 09:50:03.950700 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.000681 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6n6h\" (UniqueName: \"kubernetes.io/projected/0e0848e7-a6e0-4672-bb56-2f40193be881-kube-api-access-r6n6h\") pod \"0e0848e7-a6e0-4672-bb56-2f40193be881\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.000785 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ssh-key\") pod \"0e0848e7-a6e0-4672-bb56-2f40193be881\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.000849 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ceph\") pod \"0e0848e7-a6e0-4672-bb56-2f40193be881\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.000946 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-repo-setup-combined-ca-bundle\") pod \"0e0848e7-a6e0-4672-bb56-2f40193be881\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.001067 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-inventory\") pod \"0e0848e7-a6e0-4672-bb56-2f40193be881\" (UID: \"0e0848e7-a6e0-4672-bb56-2f40193be881\") " Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.007452 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "0e0848e7-a6e0-4672-bb56-2f40193be881" (UID: "0e0848e7-a6e0-4672-bb56-2f40193be881"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.011843 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ceph" (OuterVolumeSpecName: "ceph") pod "0e0848e7-a6e0-4672-bb56-2f40193be881" (UID: "0e0848e7-a6e0-4672-bb56-2f40193be881"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.016849 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e0848e7-a6e0-4672-bb56-2f40193be881-kube-api-access-r6n6h" (OuterVolumeSpecName: "kube-api-access-r6n6h") pod "0e0848e7-a6e0-4672-bb56-2f40193be881" (UID: "0e0848e7-a6e0-4672-bb56-2f40193be881"). InnerVolumeSpecName "kube-api-access-r6n6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.037002 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-inventory" (OuterVolumeSpecName: "inventory") pod "0e0848e7-a6e0-4672-bb56-2f40193be881" (UID: "0e0848e7-a6e0-4672-bb56-2f40193be881"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.041675 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "0e0848e7-a6e0-4672-bb56-2f40193be881" (UID: "0e0848e7-a6e0-4672-bb56-2f40193be881"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.102697 4815 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.102728 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.102739 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6n6h\" (UniqueName: \"kubernetes.io/projected/0e0848e7-a6e0-4672-bb56-2f40193be881-kube-api-access-r6n6h\") on node \"crc\" DevicePath \"\"" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.102749 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.102756 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/0e0848e7-a6e0-4672-bb56-2f40193be881-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.550187 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" event={"ID":"0e0848e7-a6e0-4672-bb56-2f40193be881","Type":"ContainerDied","Data":"70ad604d060f366fb10eaff6e335638595d54f13a688bf134b5dd01dda46bb1e"} Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.550568 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70ad604d060f366fb10eaff6e335638595d54f13a688bf134b5dd01dda46bb1e" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.550239 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.637527 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7"] Dec 05 09:50:04 crc kubenswrapper[4815]: E1205 09:50:04.638206 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e0848e7-a6e0-4672-bb56-2f40193be881" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.638305 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e0848e7-a6e0-4672-bb56-2f40193be881" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.638561 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e0848e7-a6e0-4672-bb56-2f40193be881" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.639256 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.642227 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.642272 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.642420 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.644555 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.646420 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.650026 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7"] Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.713795 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.714012 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8tfq\" (UniqueName: \"kubernetes.io/projected/296de34a-3504-4fb8-8a20-02333f447f4c-kube-api-access-k8tfq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.714060 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.714144 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.714186 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.816016 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.816350 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8tfq\" (UniqueName: \"kubernetes.io/projected/296de34a-3504-4fb8-8a20-02333f447f4c-kube-api-access-k8tfq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.816446 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.816563 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.816653 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.820374 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.820374 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.820695 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.821822 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.834032 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8tfq\" (UniqueName: \"kubernetes.io/projected/296de34a-3504-4fb8-8a20-02333f447f4c-kube-api-access-k8tfq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:04 crc kubenswrapper[4815]: I1205 09:50:04.954662 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:50:05 crc kubenswrapper[4815]: I1205 09:50:05.464457 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7"] Dec 05 09:50:05 crc kubenswrapper[4815]: I1205 09:50:05.468876 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:50:05 crc kubenswrapper[4815]: I1205 09:50:05.600186 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" event={"ID":"296de34a-3504-4fb8-8a20-02333f447f4c","Type":"ContainerStarted","Data":"3ad98e87433bc70cace790206d54e657c2b2f6bbcfdbb6cc1bb5a32169f92c1e"} Dec 05 09:50:06 crc kubenswrapper[4815]: I1205 09:50:06.610842 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" event={"ID":"296de34a-3504-4fb8-8a20-02333f447f4c","Type":"ContainerStarted","Data":"d636b74245ea6017ac984c4b06d5da6bfaf8dd3fc62c8fa0d1bef44c7dafe65b"} Dec 05 09:50:06 crc kubenswrapper[4815]: I1205 09:50:06.629248 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" podStartSLOduration=1.799868523 podStartE2EDuration="2.629228184s" podCreationTimestamp="2025-12-05 09:50:04 +0000 UTC" firstStartedPulling="2025-12-05 09:50:05.468594257 +0000 UTC m=+2604.347201084" lastFinishedPulling="2025-12-05 09:50:06.297953908 +0000 UTC m=+2605.176560745" observedRunningTime="2025-12-05 09:50:06.628234126 +0000 UTC m=+2605.506840963" watchObservedRunningTime="2025-12-05 09:50:06.629228184 +0000 UTC m=+2605.507835021" Dec 05 09:50:09 crc kubenswrapper[4815]: I1205 09:50:09.418620 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:50:09 crc kubenswrapper[4815]: E1205 09:50:09.419171 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:50:21 crc kubenswrapper[4815]: I1205 09:50:21.424597 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:50:21 crc kubenswrapper[4815]: I1205 09:50:21.759878 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"92d524fe5b80f191d3ca613cde5455e954b4de4abdbc8a66cdde8df7e0934c6c"} Dec 05 09:52:08 crc kubenswrapper[4815]: I1205 09:52:08.716972 4815 generic.go:334] "Generic (PLEG): container finished" podID="296de34a-3504-4fb8-8a20-02333f447f4c" containerID="d636b74245ea6017ac984c4b06d5da6bfaf8dd3fc62c8fa0d1bef44c7dafe65b" exitCode=0 Dec 05 09:52:08 crc kubenswrapper[4815]: I1205 09:52:08.717051 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" event={"ID":"296de34a-3504-4fb8-8a20-02333f447f4c","Type":"ContainerDied","Data":"d636b74245ea6017ac984c4b06d5da6bfaf8dd3fc62c8fa0d1bef44c7dafe65b"} Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.149192 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.272834 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-bootstrap-combined-ca-bundle\") pod \"296de34a-3504-4fb8-8a20-02333f447f4c\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.272898 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ceph\") pod \"296de34a-3504-4fb8-8a20-02333f447f4c\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.272934 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ssh-key\") pod \"296de34a-3504-4fb8-8a20-02333f447f4c\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.273046 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-inventory\") pod \"296de34a-3504-4fb8-8a20-02333f447f4c\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.273773 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8tfq\" (UniqueName: \"kubernetes.io/projected/296de34a-3504-4fb8-8a20-02333f447f4c-kube-api-access-k8tfq\") pod \"296de34a-3504-4fb8-8a20-02333f447f4c\" (UID: \"296de34a-3504-4fb8-8a20-02333f447f4c\") " Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.279155 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/296de34a-3504-4fb8-8a20-02333f447f4c-kube-api-access-k8tfq" (OuterVolumeSpecName: "kube-api-access-k8tfq") pod "296de34a-3504-4fb8-8a20-02333f447f4c" (UID: "296de34a-3504-4fb8-8a20-02333f447f4c"). InnerVolumeSpecName "kube-api-access-k8tfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.279698 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ceph" (OuterVolumeSpecName: "ceph") pod "296de34a-3504-4fb8-8a20-02333f447f4c" (UID: "296de34a-3504-4fb8-8a20-02333f447f4c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.284419 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "296de34a-3504-4fb8-8a20-02333f447f4c" (UID: "296de34a-3504-4fb8-8a20-02333f447f4c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.301024 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-inventory" (OuterVolumeSpecName: "inventory") pod "296de34a-3504-4fb8-8a20-02333f447f4c" (UID: "296de34a-3504-4fb8-8a20-02333f447f4c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.301376 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "296de34a-3504-4fb8-8a20-02333f447f4c" (UID: "296de34a-3504-4fb8-8a20-02333f447f4c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.376175 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.376213 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8tfq\" (UniqueName: \"kubernetes.io/projected/296de34a-3504-4fb8-8a20-02333f447f4c-kube-api-access-k8tfq\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.376236 4815 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.376261 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.376272 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/296de34a-3504-4fb8-8a20-02333f447f4c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.736431 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" event={"ID":"296de34a-3504-4fb8-8a20-02333f447f4c","Type":"ContainerDied","Data":"3ad98e87433bc70cace790206d54e657c2b2f6bbcfdbb6cc1bb5a32169f92c1e"} Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.736479 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ad98e87433bc70cace790206d54e657c2b2f6bbcfdbb6cc1bb5a32169f92c1e" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.736566 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.838380 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn"] Dec 05 09:52:10 crc kubenswrapper[4815]: E1205 09:52:10.838938 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="296de34a-3504-4fb8-8a20-02333f447f4c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.838975 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="296de34a-3504-4fb8-8a20-02333f447f4c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.839263 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="296de34a-3504-4fb8-8a20-02333f447f4c" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.840003 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.842040 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.846426 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.846586 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.846689 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.846893 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.854992 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn"] Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.989047 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.989578 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.989627 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:10 crc kubenswrapper[4815]: I1205 09:52:10.989660 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpwjf\" (UniqueName: \"kubernetes.io/projected/06fceca7-a988-4d36-8a56-fbe8fb96a28b-kube-api-access-tpwjf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:11 crc kubenswrapper[4815]: I1205 09:52:11.090869 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpwjf\" (UniqueName: \"kubernetes.io/projected/06fceca7-a988-4d36-8a56-fbe8fb96a28b-kube-api-access-tpwjf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:11 crc kubenswrapper[4815]: I1205 09:52:11.090988 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:11 crc kubenswrapper[4815]: I1205 09:52:11.091103 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:11 crc kubenswrapper[4815]: I1205 09:52:11.091147 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:11 crc kubenswrapper[4815]: I1205 09:52:11.097195 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:11 crc kubenswrapper[4815]: I1205 09:52:11.098620 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:11 crc kubenswrapper[4815]: I1205 09:52:11.101087 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:11 crc kubenswrapper[4815]: I1205 09:52:11.119032 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpwjf\" (UniqueName: \"kubernetes.io/projected/06fceca7-a988-4d36-8a56-fbe8fb96a28b-kube-api-access-tpwjf\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-699sn\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:11 crc kubenswrapper[4815]: I1205 09:52:11.173021 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:11 crc kubenswrapper[4815]: I1205 09:52:11.760175 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn"] Dec 05 09:52:12 crc kubenswrapper[4815]: I1205 09:52:12.753157 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" event={"ID":"06fceca7-a988-4d36-8a56-fbe8fb96a28b","Type":"ContainerStarted","Data":"d52418419c06c7e10a15564beefd2fd067ff5232a9592d29dda2175e33079993"} Dec 05 09:52:13 crc kubenswrapper[4815]: I1205 09:52:13.760699 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" event={"ID":"06fceca7-a988-4d36-8a56-fbe8fb96a28b","Type":"ContainerStarted","Data":"36d8f69af267176053aee0e4d0da31b912cfa59eafb9ce848536c3c9dd304f9f"} Dec 05 09:52:13 crc kubenswrapper[4815]: I1205 09:52:13.779613 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" podStartSLOduration=2.495512228 podStartE2EDuration="3.779586489s" podCreationTimestamp="2025-12-05 09:52:10 +0000 UTC" firstStartedPulling="2025-12-05 09:52:11.764945275 +0000 UTC m=+2730.643552112" lastFinishedPulling="2025-12-05 09:52:13.049019536 +0000 UTC m=+2731.927626373" observedRunningTime="2025-12-05 09:52:13.779063685 +0000 UTC m=+2732.657670522" watchObservedRunningTime="2025-12-05 09:52:13.779586489 +0000 UTC m=+2732.658193346" Dec 05 09:52:43 crc kubenswrapper[4815]: I1205 09:52:43.997890 4815 generic.go:334] "Generic (PLEG): container finished" podID="06fceca7-a988-4d36-8a56-fbe8fb96a28b" containerID="36d8f69af267176053aee0e4d0da31b912cfa59eafb9ce848536c3c9dd304f9f" exitCode=0 Dec 05 09:52:43 crc kubenswrapper[4815]: I1205 09:52:43.997979 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" event={"ID":"06fceca7-a988-4d36-8a56-fbe8fb96a28b","Type":"ContainerDied","Data":"36d8f69af267176053aee0e4d0da31b912cfa59eafb9ce848536c3c9dd304f9f"} Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.505091 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.656133 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpwjf\" (UniqueName: \"kubernetes.io/projected/06fceca7-a988-4d36-8a56-fbe8fb96a28b-kube-api-access-tpwjf\") pod \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.656262 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-inventory\") pod \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.656305 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ssh-key\") pod \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.656453 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ceph\") pod \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\" (UID: \"06fceca7-a988-4d36-8a56-fbe8fb96a28b\") " Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.665682 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ceph" (OuterVolumeSpecName: "ceph") pod "06fceca7-a988-4d36-8a56-fbe8fb96a28b" (UID: "06fceca7-a988-4d36-8a56-fbe8fb96a28b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.665935 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06fceca7-a988-4d36-8a56-fbe8fb96a28b-kube-api-access-tpwjf" (OuterVolumeSpecName: "kube-api-access-tpwjf") pod "06fceca7-a988-4d36-8a56-fbe8fb96a28b" (UID: "06fceca7-a988-4d36-8a56-fbe8fb96a28b"). InnerVolumeSpecName "kube-api-access-tpwjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.689803 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-inventory" (OuterVolumeSpecName: "inventory") pod "06fceca7-a988-4d36-8a56-fbe8fb96a28b" (UID: "06fceca7-a988-4d36-8a56-fbe8fb96a28b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.692687 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "06fceca7-a988-4d36-8a56-fbe8fb96a28b" (UID: "06fceca7-a988-4d36-8a56-fbe8fb96a28b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.758143 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.758869 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpwjf\" (UniqueName: \"kubernetes.io/projected/06fceca7-a988-4d36-8a56-fbe8fb96a28b-kube-api-access-tpwjf\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.758890 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:45 crc kubenswrapper[4815]: I1205 09:52:45.758898 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/06fceca7-a988-4d36-8a56-fbe8fb96a28b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.015723 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" event={"ID":"06fceca7-a988-4d36-8a56-fbe8fb96a28b","Type":"ContainerDied","Data":"d52418419c06c7e10a15564beefd2fd067ff5232a9592d29dda2175e33079993"} Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.015772 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d52418419c06c7e10a15564beefd2fd067ff5232a9592d29dda2175e33079993" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.015810 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-699sn" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.112452 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2"] Dec 05 09:52:46 crc kubenswrapper[4815]: E1205 09:52:46.113165 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="06fceca7-a988-4d36-8a56-fbe8fb96a28b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.113198 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="06fceca7-a988-4d36-8a56-fbe8fb96a28b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.113577 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="06fceca7-a988-4d36-8a56-fbe8fb96a28b" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.114902 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.118714 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.121328 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.121420 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.121581 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.122245 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.135681 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2"] Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.165989 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.166097 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq6jc\" (UniqueName: \"kubernetes.io/projected/abc71b71-d4f3-4c99-8a40-8969b4371f39-kube-api-access-tq6jc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.166188 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.166213 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.267777 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.268112 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.268265 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.268472 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq6jc\" (UniqueName: \"kubernetes.io/projected/abc71b71-d4f3-4c99-8a40-8969b4371f39-kube-api-access-tq6jc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.272300 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.276183 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.281776 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.289825 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq6jc\" (UniqueName: \"kubernetes.io/projected/abc71b71-d4f3-4c99-8a40-8969b4371f39-kube-api-access-tq6jc\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-f54t2\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.432229 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:52:46 crc kubenswrapper[4815]: I1205 09:52:46.965559 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2"] Dec 05 09:52:47 crc kubenswrapper[4815]: I1205 09:52:47.025981 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" event={"ID":"abc71b71-d4f3-4c99-8a40-8969b4371f39","Type":"ContainerStarted","Data":"6c42979b2728b909b43ca05e5e08a42acbb913f9d6eebd54984cf80bac9c5029"} Dec 05 09:52:49 crc kubenswrapper[4815]: I1205 09:52:49.764558 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x6sz8"] Dec 05 09:52:49 crc kubenswrapper[4815]: I1205 09:52:49.771600 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:52:49 crc kubenswrapper[4815]: I1205 09:52:49.797761 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6sz8"] Dec 05 09:52:49 crc kubenswrapper[4815]: I1205 09:52:49.949289 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twlll\" (UniqueName: \"kubernetes.io/projected/b0365dd2-c0ef-401b-9b01-e4607933a0fd-kube-api-access-twlll\") pod \"redhat-marketplace-x6sz8\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:52:49 crc kubenswrapper[4815]: I1205 09:52:49.949347 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-catalog-content\") pod \"redhat-marketplace-x6sz8\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:52:49 crc kubenswrapper[4815]: I1205 09:52:49.949534 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-utilities\") pod \"redhat-marketplace-x6sz8\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:52:50 crc kubenswrapper[4815]: I1205 09:52:50.050948 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-utilities\") pod \"redhat-marketplace-x6sz8\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:52:50 crc kubenswrapper[4815]: I1205 09:52:50.051046 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twlll\" (UniqueName: \"kubernetes.io/projected/b0365dd2-c0ef-401b-9b01-e4607933a0fd-kube-api-access-twlll\") pod \"redhat-marketplace-x6sz8\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:52:50 crc kubenswrapper[4815]: I1205 09:52:50.051072 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-catalog-content\") pod \"redhat-marketplace-x6sz8\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:52:50 crc kubenswrapper[4815]: I1205 09:52:50.051573 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-utilities\") pod \"redhat-marketplace-x6sz8\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:52:50 crc kubenswrapper[4815]: I1205 09:52:50.051659 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-catalog-content\") pod \"redhat-marketplace-x6sz8\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:52:50 crc kubenswrapper[4815]: I1205 09:52:50.071544 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twlll\" (UniqueName: \"kubernetes.io/projected/b0365dd2-c0ef-401b-9b01-e4607933a0fd-kube-api-access-twlll\") pod \"redhat-marketplace-x6sz8\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:52:50 crc kubenswrapper[4815]: I1205 09:52:50.101359 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:52:50 crc kubenswrapper[4815]: I1205 09:52:50.192675 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:52:50 crc kubenswrapper[4815]: I1205 09:52:50.192747 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:52:53 crc kubenswrapper[4815]: I1205 09:52:53.633462 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6sz8"] Dec 05 09:52:53 crc kubenswrapper[4815]: W1205 09:52:53.646215 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb0365dd2_c0ef_401b_9b01_e4607933a0fd.slice/crio-1d65100f85a7d3d6a1fed8fb4dfe5d84a1c22298a40b4d749b79b7f50bd081a7 WatchSource:0}: Error finding container 1d65100f85a7d3d6a1fed8fb4dfe5d84a1c22298a40b4d749b79b7f50bd081a7: Status 404 returned error can't find the container with id 1d65100f85a7d3d6a1fed8fb4dfe5d84a1c22298a40b4d749b79b7f50bd081a7 Dec 05 09:52:54 crc kubenswrapper[4815]: I1205 09:52:54.093450 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" event={"ID":"abc71b71-d4f3-4c99-8a40-8969b4371f39","Type":"ContainerStarted","Data":"3476bf57a75505a2710392e8b44f4077c22cd1dd6df0333039ba08d3a8ac0797"} Dec 05 09:52:54 crc kubenswrapper[4815]: I1205 09:52:54.098187 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6sz8" event={"ID":"b0365dd2-c0ef-401b-9b01-e4607933a0fd","Type":"ContainerStarted","Data":"1d65100f85a7d3d6a1fed8fb4dfe5d84a1c22298a40b4d749b79b7f50bd081a7"} Dec 05 09:52:54 crc kubenswrapper[4815]: I1205 09:52:54.132605 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" podStartSLOduration=1.966027566 podStartE2EDuration="8.132586507s" podCreationTimestamp="2025-12-05 09:52:46 +0000 UTC" firstStartedPulling="2025-12-05 09:52:46.973666211 +0000 UTC m=+2765.852273048" lastFinishedPulling="2025-12-05 09:52:53.140225152 +0000 UTC m=+2772.018831989" observedRunningTime="2025-12-05 09:52:54.127973546 +0000 UTC m=+2773.006580383" watchObservedRunningTime="2025-12-05 09:52:54.132586507 +0000 UTC m=+2773.011193344" Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.108527 4815 generic.go:334] "Generic (PLEG): container finished" podID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerID="72a62cbccb773cc60e40a6ea0b936734116189ae2844f88b853cbb4460d7a86b" exitCode=0 Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.108579 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6sz8" event={"ID":"b0365dd2-c0ef-401b-9b01-e4607933a0fd","Type":"ContainerDied","Data":"72a62cbccb773cc60e40a6ea0b936734116189ae2844f88b853cbb4460d7a86b"} Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.752229 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-bwvn7"] Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.754892 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.790998 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bwvn7"] Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.791452 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7czd\" (UniqueName: \"kubernetes.io/projected/4105a87a-38ab-4857-8da2-faf403753c0d-kube-api-access-r7czd\") pod \"certified-operators-bwvn7\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.792058 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-catalog-content\") pod \"certified-operators-bwvn7\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.792133 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-utilities\") pod \"certified-operators-bwvn7\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.894552 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-utilities\") pod \"certified-operators-bwvn7\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.894756 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7czd\" (UniqueName: \"kubernetes.io/projected/4105a87a-38ab-4857-8da2-faf403753c0d-kube-api-access-r7czd\") pod \"certified-operators-bwvn7\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.894830 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-catalog-content\") pod \"certified-operators-bwvn7\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.895085 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-utilities\") pod \"certified-operators-bwvn7\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.895336 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-catalog-content\") pod \"certified-operators-bwvn7\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:52:55 crc kubenswrapper[4815]: I1205 09:52:55.924652 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7czd\" (UniqueName: \"kubernetes.io/projected/4105a87a-38ab-4857-8da2-faf403753c0d-kube-api-access-r7czd\") pod \"certified-operators-bwvn7\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:52:56 crc kubenswrapper[4815]: I1205 09:52:56.117192 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6sz8" event={"ID":"b0365dd2-c0ef-401b-9b01-e4607933a0fd","Type":"ContainerStarted","Data":"8fee1af0606f41c8fffddc64821ce14b213ccd299db43d304a19cf03ab8b353f"} Dec 05 09:52:56 crc kubenswrapper[4815]: I1205 09:52:56.147102 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:52:56 crc kubenswrapper[4815]: I1205 09:52:56.901523 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-bwvn7"] Dec 05 09:52:57 crc kubenswrapper[4815]: I1205 09:52:57.130804 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvn7" event={"ID":"4105a87a-38ab-4857-8da2-faf403753c0d","Type":"ContainerStarted","Data":"a7cfa4748bea2501500211421ab5dd0955c5ca96a6e1931e98e7f8cfabc353a9"} Dec 05 09:52:58 crc kubenswrapper[4815]: I1205 09:52:58.142542 4815 generic.go:334] "Generic (PLEG): container finished" podID="4105a87a-38ab-4857-8da2-faf403753c0d" containerID="587937535e1a33a411936bbfe76391794f56ea9419e81a1016ecbdd123da7d47" exitCode=0 Dec 05 09:52:58 crc kubenswrapper[4815]: I1205 09:52:58.142912 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvn7" event={"ID":"4105a87a-38ab-4857-8da2-faf403753c0d","Type":"ContainerDied","Data":"587937535e1a33a411936bbfe76391794f56ea9419e81a1016ecbdd123da7d47"} Dec 05 09:52:58 crc kubenswrapper[4815]: I1205 09:52:58.147915 4815 generic.go:334] "Generic (PLEG): container finished" podID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerID="8fee1af0606f41c8fffddc64821ce14b213ccd299db43d304a19cf03ab8b353f" exitCode=0 Dec 05 09:52:58 crc kubenswrapper[4815]: I1205 09:52:58.147968 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6sz8" event={"ID":"b0365dd2-c0ef-401b-9b01-e4607933a0fd","Type":"ContainerDied","Data":"8fee1af0606f41c8fffddc64821ce14b213ccd299db43d304a19cf03ab8b353f"} Dec 05 09:52:59 crc kubenswrapper[4815]: I1205 09:52:59.158003 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvn7" event={"ID":"4105a87a-38ab-4857-8da2-faf403753c0d","Type":"ContainerStarted","Data":"9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6"} Dec 05 09:52:59 crc kubenswrapper[4815]: I1205 09:52:59.161074 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6sz8" event={"ID":"b0365dd2-c0ef-401b-9b01-e4607933a0fd","Type":"ContainerStarted","Data":"6ea402aadb38bd6024a1b87b5bd25e066c0d66473dd3fc8ccb98c9b64f4de8ee"} Dec 05 09:52:59 crc kubenswrapper[4815]: I1205 09:52:59.200247 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x6sz8" podStartSLOduration=6.739431473 podStartE2EDuration="10.200226641s" podCreationTimestamp="2025-12-05 09:52:49 +0000 UTC" firstStartedPulling="2025-12-05 09:52:55.110355919 +0000 UTC m=+2773.988962756" lastFinishedPulling="2025-12-05 09:52:58.571151087 +0000 UTC m=+2777.449757924" observedRunningTime="2025-12-05 09:52:59.197686079 +0000 UTC m=+2778.076292916" watchObservedRunningTime="2025-12-05 09:52:59.200226641 +0000 UTC m=+2778.078833478" Dec 05 09:53:00 crc kubenswrapper[4815]: I1205 09:53:00.102357 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:53:00 crc kubenswrapper[4815]: I1205 09:53:00.102426 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:53:00 crc kubenswrapper[4815]: I1205 09:53:00.170087 4815 generic.go:334] "Generic (PLEG): container finished" podID="abc71b71-d4f3-4c99-8a40-8969b4371f39" containerID="3476bf57a75505a2710392e8b44f4077c22cd1dd6df0333039ba08d3a8ac0797" exitCode=0 Dec 05 09:53:00 crc kubenswrapper[4815]: I1205 09:53:00.170172 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" event={"ID":"abc71b71-d4f3-4c99-8a40-8969b4371f39","Type":"ContainerDied","Data":"3476bf57a75505a2710392e8b44f4077c22cd1dd6df0333039ba08d3a8ac0797"} Dec 05 09:53:00 crc kubenswrapper[4815]: I1205 09:53:00.172581 4815 generic.go:334] "Generic (PLEG): container finished" podID="4105a87a-38ab-4857-8da2-faf403753c0d" containerID="9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6" exitCode=0 Dec 05 09:53:00 crc kubenswrapper[4815]: I1205 09:53:00.172620 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvn7" event={"ID":"4105a87a-38ab-4857-8da2-faf403753c0d","Type":"ContainerDied","Data":"9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6"} Dec 05 09:53:01 crc kubenswrapper[4815]: I1205 09:53:01.153909 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-x6sz8" podUID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerName="registry-server" probeResult="failure" output=< Dec 05 09:53:01 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 09:53:01 crc kubenswrapper[4815]: > Dec 05 09:53:01 crc kubenswrapper[4815]: I1205 09:53:01.810425 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:53:01 crc kubenswrapper[4815]: I1205 09:53:01.991359 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq6jc\" (UniqueName: \"kubernetes.io/projected/abc71b71-d4f3-4c99-8a40-8969b4371f39-kube-api-access-tq6jc\") pod \"abc71b71-d4f3-4c99-8a40-8969b4371f39\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " Dec 05 09:53:01 crc kubenswrapper[4815]: I1205 09:53:01.991418 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ssh-key\") pod \"abc71b71-d4f3-4c99-8a40-8969b4371f39\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " Dec 05 09:53:01 crc kubenswrapper[4815]: I1205 09:53:01.991525 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ceph\") pod \"abc71b71-d4f3-4c99-8a40-8969b4371f39\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " Dec 05 09:53:01 crc kubenswrapper[4815]: I1205 09:53:01.991579 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-inventory\") pod \"abc71b71-d4f3-4c99-8a40-8969b4371f39\" (UID: \"abc71b71-d4f3-4c99-8a40-8969b4371f39\") " Dec 05 09:53:01 crc kubenswrapper[4815]: I1205 09:53:01.997137 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abc71b71-d4f3-4c99-8a40-8969b4371f39-kube-api-access-tq6jc" (OuterVolumeSpecName: "kube-api-access-tq6jc") pod "abc71b71-d4f3-4c99-8a40-8969b4371f39" (UID: "abc71b71-d4f3-4c99-8a40-8969b4371f39"). InnerVolumeSpecName "kube-api-access-tq6jc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:53:01 crc kubenswrapper[4815]: I1205 09:53:01.997596 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ceph" (OuterVolumeSpecName: "ceph") pod "abc71b71-d4f3-4c99-8a40-8969b4371f39" (UID: "abc71b71-d4f3-4c99-8a40-8969b4371f39"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.024003 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "abc71b71-d4f3-4c99-8a40-8969b4371f39" (UID: "abc71b71-d4f3-4c99-8a40-8969b4371f39"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.030868 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-inventory" (OuterVolumeSpecName: "inventory") pod "abc71b71-d4f3-4c99-8a40-8969b4371f39" (UID: "abc71b71-d4f3-4c99-8a40-8969b4371f39"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.093958 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.093999 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.094012 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq6jc\" (UniqueName: \"kubernetes.io/projected/abc71b71-d4f3-4c99-8a40-8969b4371f39-kube-api-access-tq6jc\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.094023 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abc71b71-d4f3-4c99-8a40-8969b4371f39-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.199753 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" event={"ID":"abc71b71-d4f3-4c99-8a40-8969b4371f39","Type":"ContainerDied","Data":"6c42979b2728b909b43ca05e5e08a42acbb913f9d6eebd54984cf80bac9c5029"} Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.199781 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-f54t2" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.199801 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c42979b2728b909b43ca05e5e08a42acbb913f9d6eebd54984cf80bac9c5029" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.296205 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg"] Dec 05 09:53:02 crc kubenswrapper[4815]: E1205 09:53:02.296619 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abc71b71-d4f3-4c99-8a40-8969b4371f39" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.296647 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="abc71b71-d4f3-4c99-8a40-8969b4371f39" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.296872 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="abc71b71-d4f3-4c99-8a40-8969b4371f39" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.297553 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.301303 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.304685 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.305630 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.308525 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.315419 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.387542 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg"] Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.398751 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.399086 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tv4c8\" (UniqueName: \"kubernetes.io/projected/711134c0-16f2-4ba4-8120-1c39b9d833b2-kube-api-access-tv4c8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.399279 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.399346 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.501111 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.501176 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.501240 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.501347 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tv4c8\" (UniqueName: \"kubernetes.io/projected/711134c0-16f2-4ba4-8120-1c39b9d833b2-kube-api-access-tv4c8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.505516 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.506369 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.507351 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.520631 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tv4c8\" (UniqueName: \"kubernetes.io/projected/711134c0-16f2-4ba4-8120-1c39b9d833b2-kube-api-access-tv4c8\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-cx6hg\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:02 crc kubenswrapper[4815]: I1205 09:53:02.615731 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:03 crc kubenswrapper[4815]: I1205 09:53:03.174018 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg"] Dec 05 09:53:03 crc kubenswrapper[4815]: I1205 09:53:03.208797 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" event={"ID":"711134c0-16f2-4ba4-8120-1c39b9d833b2","Type":"ContainerStarted","Data":"f3fe7aa2c53a59e94c412e2e55118a0ecadbac2d11ca4ecd6ec7171fa0fdac7d"} Dec 05 09:53:08 crc kubenswrapper[4815]: I1205 09:53:08.248806 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvn7" event={"ID":"4105a87a-38ab-4857-8da2-faf403753c0d","Type":"ContainerStarted","Data":"a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481"} Dec 05 09:53:08 crc kubenswrapper[4815]: I1205 09:53:08.251422 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" event={"ID":"711134c0-16f2-4ba4-8120-1c39b9d833b2","Type":"ContainerStarted","Data":"ecd0927d0fde13cc74ec17d4bd1de4609bf61c118bc85b90d346baf39f401557"} Dec 05 09:53:08 crc kubenswrapper[4815]: I1205 09:53:08.268717 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-bwvn7" podStartSLOduration=4.892707458 podStartE2EDuration="13.268691345s" podCreationTimestamp="2025-12-05 09:52:55 +0000 UTC" firstStartedPulling="2025-12-05 09:52:58.144613948 +0000 UTC m=+2777.023220785" lastFinishedPulling="2025-12-05 09:53:06.520597835 +0000 UTC m=+2785.399204672" observedRunningTime="2025-12-05 09:53:08.268365366 +0000 UTC m=+2787.146972203" watchObservedRunningTime="2025-12-05 09:53:08.268691345 +0000 UTC m=+2787.147298182" Dec 05 09:53:09 crc kubenswrapper[4815]: I1205 09:53:09.284636 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" podStartSLOduration=3.207563962 podStartE2EDuration="7.284608867s" podCreationTimestamp="2025-12-05 09:53:02 +0000 UTC" firstStartedPulling="2025-12-05 09:53:03.75974848 +0000 UTC m=+2782.638355317" lastFinishedPulling="2025-12-05 09:53:07.836793385 +0000 UTC m=+2786.715400222" observedRunningTime="2025-12-05 09:53:09.279005923 +0000 UTC m=+2788.157612780" watchObservedRunningTime="2025-12-05 09:53:09.284608867 +0000 UTC m=+2788.163215704" Dec 05 09:53:10 crc kubenswrapper[4815]: I1205 09:53:10.152456 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:53:10 crc kubenswrapper[4815]: I1205 09:53:10.199332 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:53:10 crc kubenswrapper[4815]: I1205 09:53:10.398993 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6sz8"] Dec 05 09:53:11 crc kubenswrapper[4815]: I1205 09:53:11.274849 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x6sz8" podUID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerName="registry-server" containerID="cri-o://6ea402aadb38bd6024a1b87b5bd25e066c0d66473dd3fc8ccb98c9b64f4de8ee" gracePeriod=2 Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.147203 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.148760 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.207799 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.329398 4815 generic.go:334] "Generic (PLEG): container finished" podID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerID="6ea402aadb38bd6024a1b87b5bd25e066c0d66473dd3fc8ccb98c9b64f4de8ee" exitCode=0 Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.330368 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6sz8" event={"ID":"b0365dd2-c0ef-401b-9b01-e4607933a0fd","Type":"ContainerDied","Data":"6ea402aadb38bd6024a1b87b5bd25e066c0d66473dd3fc8ccb98c9b64f4de8ee"} Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.519433 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.694927 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.712177 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twlll\" (UniqueName: \"kubernetes.io/projected/b0365dd2-c0ef-401b-9b01-e4607933a0fd-kube-api-access-twlll\") pod \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.712280 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-catalog-content\") pod \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.712361 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-utilities\") pod \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\" (UID: \"b0365dd2-c0ef-401b-9b01-e4607933a0fd\") " Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.713524 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-utilities" (OuterVolumeSpecName: "utilities") pod "b0365dd2-c0ef-401b-9b01-e4607933a0fd" (UID: "b0365dd2-c0ef-401b-9b01-e4607933a0fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.732636 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0365dd2-c0ef-401b-9b01-e4607933a0fd-kube-api-access-twlll" (OuterVolumeSpecName: "kube-api-access-twlll") pod "b0365dd2-c0ef-401b-9b01-e4607933a0fd" (UID: "b0365dd2-c0ef-401b-9b01-e4607933a0fd"). InnerVolumeSpecName "kube-api-access-twlll". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.770559 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b0365dd2-c0ef-401b-9b01-e4607933a0fd" (UID: "b0365dd2-c0ef-401b-9b01-e4607933a0fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.814702 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twlll\" (UniqueName: \"kubernetes.io/projected/b0365dd2-c0ef-401b-9b01-e4607933a0fd-kube-api-access-twlll\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.814738 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:16 crc kubenswrapper[4815]: I1205 09:53:16.814752 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b0365dd2-c0ef-401b-9b01-e4607933a0fd-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:17 crc kubenswrapper[4815]: I1205 09:53:17.341695 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x6sz8" event={"ID":"b0365dd2-c0ef-401b-9b01-e4607933a0fd","Type":"ContainerDied","Data":"1d65100f85a7d3d6a1fed8fb4dfe5d84a1c22298a40b4d749b79b7f50bd081a7"} Dec 05 09:53:17 crc kubenswrapper[4815]: I1205 09:53:17.342460 4815 scope.go:117] "RemoveContainer" containerID="6ea402aadb38bd6024a1b87b5bd25e066c0d66473dd3fc8ccb98c9b64f4de8ee" Dec 05 09:53:17 crc kubenswrapper[4815]: I1205 09:53:17.342502 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x6sz8" Dec 05 09:53:17 crc kubenswrapper[4815]: I1205 09:53:17.375377 4815 scope.go:117] "RemoveContainer" containerID="8fee1af0606f41c8fffddc64821ce14b213ccd299db43d304a19cf03ab8b353f" Dec 05 09:53:17 crc kubenswrapper[4815]: I1205 09:53:17.382031 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6sz8"] Dec 05 09:53:17 crc kubenswrapper[4815]: I1205 09:53:17.409771 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x6sz8"] Dec 05 09:53:17 crc kubenswrapper[4815]: I1205 09:53:17.415324 4815 scope.go:117] "RemoveContainer" containerID="72a62cbccb773cc60e40a6ea0b936734116189ae2844f88b853cbb4460d7a86b" Dec 05 09:53:17 crc kubenswrapper[4815]: I1205 09:53:17.457882 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" path="/var/lib/kubelet/pods/b0365dd2-c0ef-401b-9b01-e4607933a0fd/volumes" Dec 05 09:53:18 crc kubenswrapper[4815]: I1205 09:53:18.802243 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bwvn7"] Dec 05 09:53:19 crc kubenswrapper[4815]: I1205 09:53:19.358379 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-bwvn7" podUID="4105a87a-38ab-4857-8da2-faf403753c0d" containerName="registry-server" containerID="cri-o://a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481" gracePeriod=2 Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.015222 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.016298 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7czd\" (UniqueName: \"kubernetes.io/projected/4105a87a-38ab-4857-8da2-faf403753c0d-kube-api-access-r7czd\") pod \"4105a87a-38ab-4857-8da2-faf403753c0d\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.016372 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-catalog-content\") pod \"4105a87a-38ab-4857-8da2-faf403753c0d\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.016717 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-utilities\") pod \"4105a87a-38ab-4857-8da2-faf403753c0d\" (UID: \"4105a87a-38ab-4857-8da2-faf403753c0d\") " Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.017621 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-utilities" (OuterVolumeSpecName: "utilities") pod "4105a87a-38ab-4857-8da2-faf403753c0d" (UID: "4105a87a-38ab-4857-8da2-faf403753c0d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.018140 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.020979 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4105a87a-38ab-4857-8da2-faf403753c0d-kube-api-access-r7czd" (OuterVolumeSpecName: "kube-api-access-r7czd") pod "4105a87a-38ab-4857-8da2-faf403753c0d" (UID: "4105a87a-38ab-4857-8da2-faf403753c0d"). InnerVolumeSpecName "kube-api-access-r7czd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.086040 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4105a87a-38ab-4857-8da2-faf403753c0d" (UID: "4105a87a-38ab-4857-8da2-faf403753c0d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.119763 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7czd\" (UniqueName: \"kubernetes.io/projected/4105a87a-38ab-4857-8da2-faf403753c0d-kube-api-access-r7czd\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.119795 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4105a87a-38ab-4857-8da2-faf403753c0d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.192393 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.192461 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.379027 4815 generic.go:334] "Generic (PLEG): container finished" podID="4105a87a-38ab-4857-8da2-faf403753c0d" containerID="a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481" exitCode=0 Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.379103 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-bwvn7" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.379090 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvn7" event={"ID":"4105a87a-38ab-4857-8da2-faf403753c0d","Type":"ContainerDied","Data":"a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481"} Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.379281 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-bwvn7" event={"ID":"4105a87a-38ab-4857-8da2-faf403753c0d","Type":"ContainerDied","Data":"a7cfa4748bea2501500211421ab5dd0955c5ca96a6e1931e98e7f8cfabc353a9"} Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.379308 4815 scope.go:117] "RemoveContainer" containerID="a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.418899 4815 scope.go:117] "RemoveContainer" containerID="9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.433212 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-bwvn7"] Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.435920 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-bwvn7"] Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.447731 4815 scope.go:117] "RemoveContainer" containerID="587937535e1a33a411936bbfe76391794f56ea9419e81a1016ecbdd123da7d47" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.493314 4815 scope.go:117] "RemoveContainer" containerID="a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481" Dec 05 09:53:20 crc kubenswrapper[4815]: E1205 09:53:20.493995 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481\": container with ID starting with a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481 not found: ID does not exist" containerID="a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.494040 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481"} err="failed to get container status \"a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481\": rpc error: code = NotFound desc = could not find container \"a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481\": container with ID starting with a864f5580ec5f90f508ffebdbc5a7620297c73c5b869dd7dd6d0b0b246246481 not found: ID does not exist" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.494071 4815 scope.go:117] "RemoveContainer" containerID="9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6" Dec 05 09:53:20 crc kubenswrapper[4815]: E1205 09:53:20.494683 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6\": container with ID starting with 9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6 not found: ID does not exist" containerID="9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.494717 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6"} err="failed to get container status \"9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6\": rpc error: code = NotFound desc = could not find container \"9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6\": container with ID starting with 9c53838769806ae3c6873974a88a3cc30f86958c9ba73ede3b14333149bde5d6 not found: ID does not exist" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.494741 4815 scope.go:117] "RemoveContainer" containerID="587937535e1a33a411936bbfe76391794f56ea9419e81a1016ecbdd123da7d47" Dec 05 09:53:20 crc kubenswrapper[4815]: E1205 09:53:20.495021 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"587937535e1a33a411936bbfe76391794f56ea9419e81a1016ecbdd123da7d47\": container with ID starting with 587937535e1a33a411936bbfe76391794f56ea9419e81a1016ecbdd123da7d47 not found: ID does not exist" containerID="587937535e1a33a411936bbfe76391794f56ea9419e81a1016ecbdd123da7d47" Dec 05 09:53:20 crc kubenswrapper[4815]: I1205 09:53:20.495042 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"587937535e1a33a411936bbfe76391794f56ea9419e81a1016ecbdd123da7d47"} err="failed to get container status \"587937535e1a33a411936bbfe76391794f56ea9419e81a1016ecbdd123da7d47\": rpc error: code = NotFound desc = could not find container \"587937535e1a33a411936bbfe76391794f56ea9419e81a1016ecbdd123da7d47\": container with ID starting with 587937535e1a33a411936bbfe76391794f56ea9419e81a1016ecbdd123da7d47 not found: ID does not exist" Dec 05 09:53:21 crc kubenswrapper[4815]: I1205 09:53:21.429897 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4105a87a-38ab-4857-8da2-faf403753c0d" path="/var/lib/kubelet/pods/4105a87a-38ab-4857-8da2-faf403753c0d/volumes" Dec 05 09:53:50 crc kubenswrapper[4815]: I1205 09:53:50.192941 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:53:50 crc kubenswrapper[4815]: I1205 09:53:50.193814 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:53:50 crc kubenswrapper[4815]: I1205 09:53:50.193911 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:53:50 crc kubenswrapper[4815]: I1205 09:53:50.195059 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"92d524fe5b80f191d3ca613cde5455e954b4de4abdbc8a66cdde8df7e0934c6c"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:53:50 crc kubenswrapper[4815]: I1205 09:53:50.195162 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://92d524fe5b80f191d3ca613cde5455e954b4de4abdbc8a66cdde8df7e0934c6c" gracePeriod=600 Dec 05 09:53:50 crc kubenswrapper[4815]: I1205 09:53:50.684219 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="92d524fe5b80f191d3ca613cde5455e954b4de4abdbc8a66cdde8df7e0934c6c" exitCode=0 Dec 05 09:53:50 crc kubenswrapper[4815]: I1205 09:53:50.684254 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"92d524fe5b80f191d3ca613cde5455e954b4de4abdbc8a66cdde8df7e0934c6c"} Dec 05 09:53:50 crc kubenswrapper[4815]: I1205 09:53:50.684622 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754"} Dec 05 09:53:50 crc kubenswrapper[4815]: I1205 09:53:50.684702 4815 scope.go:117] "RemoveContainer" containerID="3997367ee75503e0a5fcbc7021857e1a4ad1b8c4cd0cff0cdfdae60416084372" Dec 05 09:53:55 crc kubenswrapper[4815]: I1205 09:53:55.756268 4815 generic.go:334] "Generic (PLEG): container finished" podID="711134c0-16f2-4ba4-8120-1c39b9d833b2" containerID="ecd0927d0fde13cc74ec17d4bd1de4609bf61c118bc85b90d346baf39f401557" exitCode=0 Dec 05 09:53:55 crc kubenswrapper[4815]: I1205 09:53:55.756801 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" event={"ID":"711134c0-16f2-4ba4-8120-1c39b9d833b2","Type":"ContainerDied","Data":"ecd0927d0fde13cc74ec17d4bd1de4609bf61c118bc85b90d346baf39f401557"} Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.179341 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.258890 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ssh-key\") pod \"711134c0-16f2-4ba4-8120-1c39b9d833b2\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.258933 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tv4c8\" (UniqueName: \"kubernetes.io/projected/711134c0-16f2-4ba4-8120-1c39b9d833b2-kube-api-access-tv4c8\") pod \"711134c0-16f2-4ba4-8120-1c39b9d833b2\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.258999 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ceph\") pod \"711134c0-16f2-4ba4-8120-1c39b9d833b2\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.259088 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-inventory\") pod \"711134c0-16f2-4ba4-8120-1c39b9d833b2\" (UID: \"711134c0-16f2-4ba4-8120-1c39b9d833b2\") " Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.273154 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ceph" (OuterVolumeSpecName: "ceph") pod "711134c0-16f2-4ba4-8120-1c39b9d833b2" (UID: "711134c0-16f2-4ba4-8120-1c39b9d833b2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.280792 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/711134c0-16f2-4ba4-8120-1c39b9d833b2-kube-api-access-tv4c8" (OuterVolumeSpecName: "kube-api-access-tv4c8") pod "711134c0-16f2-4ba4-8120-1c39b9d833b2" (UID: "711134c0-16f2-4ba4-8120-1c39b9d833b2"). InnerVolumeSpecName "kube-api-access-tv4c8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.306357 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-inventory" (OuterVolumeSpecName: "inventory") pod "711134c0-16f2-4ba4-8120-1c39b9d833b2" (UID: "711134c0-16f2-4ba4-8120-1c39b9d833b2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.306400 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "711134c0-16f2-4ba4-8120-1c39b9d833b2" (UID: "711134c0-16f2-4ba4-8120-1c39b9d833b2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.360346 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.360390 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.360408 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/711134c0-16f2-4ba4-8120-1c39b9d833b2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.360424 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tv4c8\" (UniqueName: \"kubernetes.io/projected/711134c0-16f2-4ba4-8120-1c39b9d833b2-kube-api-access-tv4c8\") on node \"crc\" DevicePath \"\"" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.775246 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" event={"ID":"711134c0-16f2-4ba4-8120-1c39b9d833b2","Type":"ContainerDied","Data":"f3fe7aa2c53a59e94c412e2e55118a0ecadbac2d11ca4ecd6ec7171fa0fdac7d"} Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.775289 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3fe7aa2c53a59e94c412e2e55118a0ecadbac2d11ca4ecd6ec7171fa0fdac7d" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.775740 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-cx6hg" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.927154 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r"] Dec 05 09:53:57 crc kubenswrapper[4815]: E1205 09:53:57.927557 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerName="extract-utilities" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.927586 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerName="extract-utilities" Dec 05 09:53:57 crc kubenswrapper[4815]: E1205 09:53:57.927600 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4105a87a-38ab-4857-8da2-faf403753c0d" containerName="extract-utilities" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.927606 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4105a87a-38ab-4857-8da2-faf403753c0d" containerName="extract-utilities" Dec 05 09:53:57 crc kubenswrapper[4815]: E1205 09:53:57.927615 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerName="registry-server" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.927621 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerName="registry-server" Dec 05 09:53:57 crc kubenswrapper[4815]: E1205 09:53:57.927652 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="711134c0-16f2-4ba4-8120-1c39b9d833b2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.927660 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="711134c0-16f2-4ba4-8120-1c39b9d833b2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:53:57 crc kubenswrapper[4815]: E1205 09:53:57.927671 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerName="extract-content" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.927678 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerName="extract-content" Dec 05 09:53:57 crc kubenswrapper[4815]: E1205 09:53:57.927692 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4105a87a-38ab-4857-8da2-faf403753c0d" containerName="extract-content" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.927697 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4105a87a-38ab-4857-8da2-faf403753c0d" containerName="extract-content" Dec 05 09:53:57 crc kubenswrapper[4815]: E1205 09:53:57.927711 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4105a87a-38ab-4857-8da2-faf403753c0d" containerName="registry-server" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.927716 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4105a87a-38ab-4857-8da2-faf403753c0d" containerName="registry-server" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.927974 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="711134c0-16f2-4ba4-8120-1c39b9d833b2" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.927989 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="4105a87a-38ab-4857-8da2-faf403753c0d" containerName="registry-server" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.928003 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0365dd2-c0ef-401b-9b01-e4607933a0fd" containerName="registry-server" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.928661 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.935648 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r"] Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.936293 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.936333 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.937005 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.937134 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:53:57 crc kubenswrapper[4815]: I1205 09:53:57.937238 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.097672 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.097756 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqqc7\" (UniqueName: \"kubernetes.io/projected/18189644-5dd2-4329-a10b-70d62c5e0fd0-kube-api-access-hqqc7\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.097807 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.097884 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.199675 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.199765 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqqc7\" (UniqueName: \"kubernetes.io/projected/18189644-5dd2-4329-a10b-70d62c5e0fd0-kube-api-access-hqqc7\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.199818 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.199843 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.203395 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.203569 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.203928 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.217597 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqqc7\" (UniqueName: \"kubernetes.io/projected/18189644-5dd2-4329-a10b-70d62c5e0fd0-kube-api-access-hqqc7\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.245593 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:53:58 crc kubenswrapper[4815]: I1205 09:53:58.792978 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r"] Dec 05 09:53:59 crc kubenswrapper[4815]: I1205 09:53:59.792756 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" event={"ID":"18189644-5dd2-4329-a10b-70d62c5e0fd0","Type":"ContainerStarted","Data":"799619a46d62aef729ba997bb327e28cbb151bff8d90e9dba6123bdebe17da4a"} Dec 05 09:53:59 crc kubenswrapper[4815]: I1205 09:53:59.793836 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" event={"ID":"18189644-5dd2-4329-a10b-70d62c5e0fd0","Type":"ContainerStarted","Data":"0fb916d502546014e7bd785905ce3ddc843100524d5af8b30bfd147f135560ef"} Dec 05 09:53:59 crc kubenswrapper[4815]: I1205 09:53:59.820742 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" podStartSLOduration=2.402582962 podStartE2EDuration="2.820710123s" podCreationTimestamp="2025-12-05 09:53:57 +0000 UTC" firstStartedPulling="2025-12-05 09:53:58.822311662 +0000 UTC m=+2837.700918499" lastFinishedPulling="2025-12-05 09:53:59.240438813 +0000 UTC m=+2838.119045660" observedRunningTime="2025-12-05 09:53:59.810919014 +0000 UTC m=+2838.689525871" watchObservedRunningTime="2025-12-05 09:53:59.820710123 +0000 UTC m=+2838.699316960" Dec 05 09:54:03 crc kubenswrapper[4815]: I1205 09:54:03.825610 4815 generic.go:334] "Generic (PLEG): container finished" podID="18189644-5dd2-4329-a10b-70d62c5e0fd0" containerID="799619a46d62aef729ba997bb327e28cbb151bff8d90e9dba6123bdebe17da4a" exitCode=0 Dec 05 09:54:03 crc kubenswrapper[4815]: I1205 09:54:03.825702 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" event={"ID":"18189644-5dd2-4329-a10b-70d62c5e0fd0","Type":"ContainerDied","Data":"799619a46d62aef729ba997bb327e28cbb151bff8d90e9dba6123bdebe17da4a"} Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.250186 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.446533 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-inventory\") pod \"18189644-5dd2-4329-a10b-70d62c5e0fd0\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.446602 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hqqc7\" (UniqueName: \"kubernetes.io/projected/18189644-5dd2-4329-a10b-70d62c5e0fd0-kube-api-access-hqqc7\") pod \"18189644-5dd2-4329-a10b-70d62c5e0fd0\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.447261 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ssh-key\") pod \"18189644-5dd2-4329-a10b-70d62c5e0fd0\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.447334 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ceph\") pod \"18189644-5dd2-4329-a10b-70d62c5e0fd0\" (UID: \"18189644-5dd2-4329-a10b-70d62c5e0fd0\") " Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.452189 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18189644-5dd2-4329-a10b-70d62c5e0fd0-kube-api-access-hqqc7" (OuterVolumeSpecName: "kube-api-access-hqqc7") pod "18189644-5dd2-4329-a10b-70d62c5e0fd0" (UID: "18189644-5dd2-4329-a10b-70d62c5e0fd0"). InnerVolumeSpecName "kube-api-access-hqqc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.452372 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ceph" (OuterVolumeSpecName: "ceph") pod "18189644-5dd2-4329-a10b-70d62c5e0fd0" (UID: "18189644-5dd2-4329-a10b-70d62c5e0fd0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.473055 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-inventory" (OuterVolumeSpecName: "inventory") pod "18189644-5dd2-4329-a10b-70d62c5e0fd0" (UID: "18189644-5dd2-4329-a10b-70d62c5e0fd0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.475992 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "18189644-5dd2-4329-a10b-70d62c5e0fd0" (UID: "18189644-5dd2-4329-a10b-70d62c5e0fd0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.549109 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.549145 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.549157 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/18189644-5dd2-4329-a10b-70d62c5e0fd0-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.549167 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hqqc7\" (UniqueName: \"kubernetes.io/projected/18189644-5dd2-4329-a10b-70d62c5e0fd0-kube-api-access-hqqc7\") on node \"crc\" DevicePath \"\"" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.843937 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" event={"ID":"18189644-5dd2-4329-a10b-70d62c5e0fd0","Type":"ContainerDied","Data":"0fb916d502546014e7bd785905ce3ddc843100524d5af8b30bfd147f135560ef"} Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.844385 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fb916d502546014e7bd785905ce3ddc843100524d5af8b30bfd147f135560ef" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.844031 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.925578 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh"] Dec 05 09:54:05 crc kubenswrapper[4815]: E1205 09:54:05.926232 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18189644-5dd2-4329-a10b-70d62c5e0fd0" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.926259 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="18189644-5dd2-4329-a10b-70d62c5e0fd0" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.926515 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="18189644-5dd2-4329-a10b-70d62c5e0fd0" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.927361 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.930321 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.930550 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.930724 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.930977 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.931101 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:54:05 crc kubenswrapper[4815]: I1205 09:54:05.939827 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh"] Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.060891 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.060951 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.061048 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.061325 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gfhq\" (UniqueName: \"kubernetes.io/projected/4ab36e54-e39b-4b61-8609-52004d810c7a-kube-api-access-6gfhq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.162985 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.163119 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.163205 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gfhq\" (UniqueName: \"kubernetes.io/projected/4ab36e54-e39b-4b61-8609-52004d810c7a-kube-api-access-6gfhq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.163259 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.173603 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.173836 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.177150 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.179623 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gfhq\" (UniqueName: \"kubernetes.io/projected/4ab36e54-e39b-4b61-8609-52004d810c7a-kube-api-access-6gfhq\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.245286 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:06 crc kubenswrapper[4815]: I1205 09:54:06.946137 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh"] Dec 05 09:54:07 crc kubenswrapper[4815]: I1205 09:54:07.860799 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" event={"ID":"4ab36e54-e39b-4b61-8609-52004d810c7a","Type":"ContainerStarted","Data":"187ddb1b340ad4d4960422cd26a2411ca7fa9aa0926c8d3816b43e2fde131e01"} Dec 05 09:54:07 crc kubenswrapper[4815]: I1205 09:54:07.861131 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" event={"ID":"4ab36e54-e39b-4b61-8609-52004d810c7a","Type":"ContainerStarted","Data":"3b93430587ebbf5465136441436bd6d3d871df99693bc12836e55e42374c570c"} Dec 05 09:54:55 crc kubenswrapper[4815]: I1205 09:54:55.280538 4815 generic.go:334] "Generic (PLEG): container finished" podID="4ab36e54-e39b-4b61-8609-52004d810c7a" containerID="187ddb1b340ad4d4960422cd26a2411ca7fa9aa0926c8d3816b43e2fde131e01" exitCode=0 Dec 05 09:54:55 crc kubenswrapper[4815]: I1205 09:54:55.280576 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" event={"ID":"4ab36e54-e39b-4b61-8609-52004d810c7a","Type":"ContainerDied","Data":"187ddb1b340ad4d4960422cd26a2411ca7fa9aa0926c8d3816b43e2fde131e01"} Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.646571 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.733862 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-inventory\") pod \"4ab36e54-e39b-4b61-8609-52004d810c7a\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.733910 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ceph\") pod \"4ab36e54-e39b-4b61-8609-52004d810c7a\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.733929 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ssh-key\") pod \"4ab36e54-e39b-4b61-8609-52004d810c7a\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.734452 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gfhq\" (UniqueName: \"kubernetes.io/projected/4ab36e54-e39b-4b61-8609-52004d810c7a-kube-api-access-6gfhq\") pod \"4ab36e54-e39b-4b61-8609-52004d810c7a\" (UID: \"4ab36e54-e39b-4b61-8609-52004d810c7a\") " Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.748760 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ceph" (OuterVolumeSpecName: "ceph") pod "4ab36e54-e39b-4b61-8609-52004d810c7a" (UID: "4ab36e54-e39b-4b61-8609-52004d810c7a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.753003 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ab36e54-e39b-4b61-8609-52004d810c7a-kube-api-access-6gfhq" (OuterVolumeSpecName: "kube-api-access-6gfhq") pod "4ab36e54-e39b-4b61-8609-52004d810c7a" (UID: "4ab36e54-e39b-4b61-8609-52004d810c7a"). InnerVolumeSpecName "kube-api-access-6gfhq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.760043 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-inventory" (OuterVolumeSpecName: "inventory") pod "4ab36e54-e39b-4b61-8609-52004d810c7a" (UID: "4ab36e54-e39b-4b61-8609-52004d810c7a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.767733 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4ab36e54-e39b-4b61-8609-52004d810c7a" (UID: "4ab36e54-e39b-4b61-8609-52004d810c7a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.837080 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gfhq\" (UniqueName: \"kubernetes.io/projected/4ab36e54-e39b-4b61-8609-52004d810c7a-kube-api-access-6gfhq\") on node \"crc\" DevicePath \"\"" Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.837114 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.837123 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:54:56 crc kubenswrapper[4815]: I1205 09:54:56.837131 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4ab36e54-e39b-4b61-8609-52004d810c7a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.297578 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" event={"ID":"4ab36e54-e39b-4b61-8609-52004d810c7a","Type":"ContainerDied","Data":"3b93430587ebbf5465136441436bd6d3d871df99693bc12836e55e42374c570c"} Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.297905 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b93430587ebbf5465136441436bd6d3d871df99693bc12836e55e42374c570c" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.297624 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.393829 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-w8b66"] Dec 05 09:54:57 crc kubenswrapper[4815]: E1205 09:54:57.394370 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ab36e54-e39b-4b61-8609-52004d810c7a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.394394 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ab36e54-e39b-4b61-8609-52004d810c7a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.394644 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ab36e54-e39b-4b61-8609-52004d810c7a" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.395431 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.397697 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.397876 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.399459 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.400206 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.402712 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.413328 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-w8b66"] Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.448765 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.448844 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.448943 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ceph\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.448970 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6pl9r\" (UniqueName: \"kubernetes.io/projected/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-kube-api-access-6pl9r\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.551311 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.551745 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.551886 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ceph\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.551970 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6pl9r\" (UniqueName: \"kubernetes.io/projected/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-kube-api-access-6pl9r\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.557887 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.559274 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.562014 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ceph\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.571327 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6pl9r\" (UniqueName: \"kubernetes.io/projected/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-kube-api-access-6pl9r\") pod \"ssh-known-hosts-edpm-deployment-w8b66\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:57 crc kubenswrapper[4815]: I1205 09:54:57.714875 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:54:58 crc kubenswrapper[4815]: I1205 09:54:58.238735 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-w8b66"] Dec 05 09:54:58 crc kubenswrapper[4815]: I1205 09:54:58.305852 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" event={"ID":"ba6c4e35-e275-4cd3-8f30-0102966dcc1b","Type":"ContainerStarted","Data":"357ee768cfe3b8c120b7ce58936867ad8371aa6e11d944b11a7a68619885c718"} Dec 05 09:54:59 crc kubenswrapper[4815]: I1205 09:54:59.324720 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" event={"ID":"ba6c4e35-e275-4cd3-8f30-0102966dcc1b","Type":"ContainerStarted","Data":"0b4561b7241b666285fa4c48537a349a40cb51419acdfdcaeba123c9efcf82df"} Dec 05 09:54:59 crc kubenswrapper[4815]: I1205 09:54:59.358389 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" podStartSLOduration=1.682842129 podStartE2EDuration="2.358366973s" podCreationTimestamp="2025-12-05 09:54:57 +0000 UTC" firstStartedPulling="2025-12-05 09:54:58.243023833 +0000 UTC m=+2897.121630670" lastFinishedPulling="2025-12-05 09:54:58.918548677 +0000 UTC m=+2897.797155514" observedRunningTime="2025-12-05 09:54:59.34878397 +0000 UTC m=+2898.227390797" watchObservedRunningTime="2025-12-05 09:54:59.358366973 +0000 UTC m=+2898.236973810" Dec 05 09:55:09 crc kubenswrapper[4815]: I1205 09:55:09.408406 4815 generic.go:334] "Generic (PLEG): container finished" podID="ba6c4e35-e275-4cd3-8f30-0102966dcc1b" containerID="0b4561b7241b666285fa4c48537a349a40cb51419acdfdcaeba123c9efcf82df" exitCode=0 Dec 05 09:55:09 crc kubenswrapper[4815]: I1205 09:55:09.408518 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" event={"ID":"ba6c4e35-e275-4cd3-8f30-0102966dcc1b","Type":"ContainerDied","Data":"0b4561b7241b666285fa4c48537a349a40cb51419acdfdcaeba123c9efcf82df"} Dec 05 09:55:10 crc kubenswrapper[4815]: I1205 09:55:10.837704 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:55:10 crc kubenswrapper[4815]: I1205 09:55:10.907447 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-inventory-0\") pod \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " Dec 05 09:55:10 crc kubenswrapper[4815]: I1205 09:55:10.907647 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ceph\") pod \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " Dec 05 09:55:10 crc kubenswrapper[4815]: I1205 09:55:10.907705 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ssh-key-openstack-edpm-ipam\") pod \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " Dec 05 09:55:10 crc kubenswrapper[4815]: I1205 09:55:10.907840 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6pl9r\" (UniqueName: \"kubernetes.io/projected/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-kube-api-access-6pl9r\") pod \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\" (UID: \"ba6c4e35-e275-4cd3-8f30-0102966dcc1b\") " Dec 05 09:55:10 crc kubenswrapper[4815]: I1205 09:55:10.914009 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-kube-api-access-6pl9r" (OuterVolumeSpecName: "kube-api-access-6pl9r") pod "ba6c4e35-e275-4cd3-8f30-0102966dcc1b" (UID: "ba6c4e35-e275-4cd3-8f30-0102966dcc1b"). InnerVolumeSpecName "kube-api-access-6pl9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:55:10 crc kubenswrapper[4815]: I1205 09:55:10.919302 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ceph" (OuterVolumeSpecName: "ceph") pod "ba6c4e35-e275-4cd3-8f30-0102966dcc1b" (UID: "ba6c4e35-e275-4cd3-8f30-0102966dcc1b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:55:10 crc kubenswrapper[4815]: I1205 09:55:10.937619 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ba6c4e35-e275-4cd3-8f30-0102966dcc1b" (UID: "ba6c4e35-e275-4cd3-8f30-0102966dcc1b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:55:10 crc kubenswrapper[4815]: I1205 09:55:10.944322 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "ba6c4e35-e275-4cd3-8f30-0102966dcc1b" (UID: "ba6c4e35-e275-4cd3-8f30-0102966dcc1b"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.010785 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.010830 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.010840 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6pl9r\" (UniqueName: \"kubernetes.io/projected/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-kube-api-access-6pl9r\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.010851 4815 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/ba6c4e35-e275-4cd3-8f30-0102966dcc1b-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.429333 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.432753 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-w8b66" event={"ID":"ba6c4e35-e275-4cd3-8f30-0102966dcc1b","Type":"ContainerDied","Data":"357ee768cfe3b8c120b7ce58936867ad8371aa6e11d944b11a7a68619885c718"} Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.432788 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="357ee768cfe3b8c120b7ce58936867ad8371aa6e11d944b11a7a68619885c718" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.523158 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt"] Dec 05 09:55:11 crc kubenswrapper[4815]: E1205 09:55:11.523669 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba6c4e35-e275-4cd3-8f30-0102966dcc1b" containerName="ssh-known-hosts-edpm-deployment" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.523690 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba6c4e35-e275-4cd3-8f30-0102966dcc1b" containerName="ssh-known-hosts-edpm-deployment" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.523951 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba6c4e35-e275-4cd3-8f30-0102966dcc1b" containerName="ssh-known-hosts-edpm-deployment" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.524860 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.528729 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.529349 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.529649 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.529834 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.532790 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.545828 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt"] Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.621024 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.621086 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.621119 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hgzm\" (UniqueName: \"kubernetes.io/projected/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-kube-api-access-2hgzm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.621302 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.722814 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.722864 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hgzm\" (UniqueName: \"kubernetes.io/projected/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-kube-api-access-2hgzm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.722967 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.723050 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.737297 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.737405 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.738837 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.740959 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hgzm\" (UniqueName: \"kubernetes.io/projected/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-kube-api-access-2hgzm\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-46rkt\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:11 crc kubenswrapper[4815]: I1205 09:55:11.855216 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:12 crc kubenswrapper[4815]: I1205 09:55:12.382718 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt"] Dec 05 09:55:12 crc kubenswrapper[4815]: I1205 09:55:12.384539 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 09:55:12 crc kubenswrapper[4815]: I1205 09:55:12.456908 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" event={"ID":"7251f2d6-ee27-4b5d-9d1a-6aeab813638b","Type":"ContainerStarted","Data":"61b753da6dac7ff1ed88329e2a572363b7c1694c8cf8c198880966f29abb649a"} Dec 05 09:55:14 crc kubenswrapper[4815]: I1205 09:55:14.473590 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" event={"ID":"7251f2d6-ee27-4b5d-9d1a-6aeab813638b","Type":"ContainerStarted","Data":"6b88798f47abadeccee27a15061d3711650d9d317a682183c773c9c2a438546e"} Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.275937 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" podStartSLOduration=10.229039484 podStartE2EDuration="11.275915315s" podCreationTimestamp="2025-12-05 09:55:11 +0000 UTC" firstStartedPulling="2025-12-05 09:55:12.384213298 +0000 UTC m=+2911.262820135" lastFinishedPulling="2025-12-05 09:55:13.431089129 +0000 UTC m=+2912.309695966" observedRunningTime="2025-12-05 09:55:14.490794618 +0000 UTC m=+2913.369401465" watchObservedRunningTime="2025-12-05 09:55:22.275915315 +0000 UTC m=+2921.154522152" Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.299990 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pbw8b"] Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.302878 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.324052 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pbw8b"] Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.481858 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-catalog-content\") pod \"community-operators-pbw8b\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.481960 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-utilities\") pod \"community-operators-pbw8b\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.481994 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-268bv\" (UniqueName: \"kubernetes.io/projected/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-kube-api-access-268bv\") pod \"community-operators-pbw8b\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.559113 4815 generic.go:334] "Generic (PLEG): container finished" podID="7251f2d6-ee27-4b5d-9d1a-6aeab813638b" containerID="6b88798f47abadeccee27a15061d3711650d9d317a682183c773c9c2a438546e" exitCode=0 Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.559154 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" event={"ID":"7251f2d6-ee27-4b5d-9d1a-6aeab813638b","Type":"ContainerDied","Data":"6b88798f47abadeccee27a15061d3711650d9d317a682183c773c9c2a438546e"} Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.583675 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-catalog-content\") pod \"community-operators-pbw8b\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.583750 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-utilities\") pod \"community-operators-pbw8b\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.583778 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-268bv\" (UniqueName: \"kubernetes.io/projected/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-kube-api-access-268bv\") pod \"community-operators-pbw8b\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.584478 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-catalog-content\") pod \"community-operators-pbw8b\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.584508 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-utilities\") pod \"community-operators-pbw8b\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.616010 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-268bv\" (UniqueName: \"kubernetes.io/projected/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-kube-api-access-268bv\") pod \"community-operators-pbw8b\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:22 crc kubenswrapper[4815]: I1205 09:55:22.703259 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:23 crc kubenswrapper[4815]: I1205 09:55:23.065822 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pbw8b"] Dec 05 09:55:23 crc kubenswrapper[4815]: I1205 09:55:23.569558 4815 generic.go:334] "Generic (PLEG): container finished" podID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" containerID="0ad30ebb4a059d7315bd85832102f25a9d54122e34d0601dfb1be9899dbb434f" exitCode=0 Dec 05 09:55:23 crc kubenswrapper[4815]: I1205 09:55:23.570799 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pbw8b" event={"ID":"d0b9fc78-4e68-43b4-83c8-9d1b703b834e","Type":"ContainerDied","Data":"0ad30ebb4a059d7315bd85832102f25a9d54122e34d0601dfb1be9899dbb434f"} Dec 05 09:55:23 crc kubenswrapper[4815]: I1205 09:55:23.570831 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pbw8b" event={"ID":"d0b9fc78-4e68-43b4-83c8-9d1b703b834e","Type":"ContainerStarted","Data":"ff2d287a7ac100b74506a2190f61eadd83cfc80dd7971e574db0bcf6b301cee5"} Dec 05 09:55:23 crc kubenswrapper[4815]: I1205 09:55:23.987570 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.116697 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ceph\") pod \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.116740 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ssh-key\") pod \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.116787 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hgzm\" (UniqueName: \"kubernetes.io/projected/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-kube-api-access-2hgzm\") pod \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.116820 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-inventory\") pod \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\" (UID: \"7251f2d6-ee27-4b5d-9d1a-6aeab813638b\") " Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.124876 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-kube-api-access-2hgzm" (OuterVolumeSpecName: "kube-api-access-2hgzm") pod "7251f2d6-ee27-4b5d-9d1a-6aeab813638b" (UID: "7251f2d6-ee27-4b5d-9d1a-6aeab813638b"). InnerVolumeSpecName "kube-api-access-2hgzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.125620 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ceph" (OuterVolumeSpecName: "ceph") pod "7251f2d6-ee27-4b5d-9d1a-6aeab813638b" (UID: "7251f2d6-ee27-4b5d-9d1a-6aeab813638b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.217862 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-inventory" (OuterVolumeSpecName: "inventory") pod "7251f2d6-ee27-4b5d-9d1a-6aeab813638b" (UID: "7251f2d6-ee27-4b5d-9d1a-6aeab813638b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.219347 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.219387 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hgzm\" (UniqueName: \"kubernetes.io/projected/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-kube-api-access-2hgzm\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.219400 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.223176 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7251f2d6-ee27-4b5d-9d1a-6aeab813638b" (UID: "7251f2d6-ee27-4b5d-9d1a-6aeab813638b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.320466 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7251f2d6-ee27-4b5d-9d1a-6aeab813638b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.579113 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pbw8b" event={"ID":"d0b9fc78-4e68-43b4-83c8-9d1b703b834e","Type":"ContainerStarted","Data":"e04cd08e9eb6bc1dfd4063a5defcdeac81e3adbad58c07079ee0b9a6986426ce"} Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.580711 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" event={"ID":"7251f2d6-ee27-4b5d-9d1a-6aeab813638b","Type":"ContainerDied","Data":"61b753da6dac7ff1ed88329e2a572363b7c1694c8cf8c198880966f29abb649a"} Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.580755 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="61b753da6dac7ff1ed88329e2a572363b7c1694c8cf8c198880966f29abb649a" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.580808 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-46rkt" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.668586 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz"] Dec 05 09:55:24 crc kubenswrapper[4815]: E1205 09:55:24.668906 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7251f2d6-ee27-4b5d-9d1a-6aeab813638b" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.668923 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7251f2d6-ee27-4b5d-9d1a-6aeab813638b" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.669079 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7251f2d6-ee27-4b5d-9d1a-6aeab813638b" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.669824 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.671826 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.671985 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.672408 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.672589 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.673143 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.686113 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz"] Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.734837 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jh9f\" (UniqueName: \"kubernetes.io/projected/e775d3b4-8e2f-4255-8175-9664129b7998-kube-api-access-7jh9f\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.734939 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.734973 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.735026 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.836242 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.836319 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.836394 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.836538 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jh9f\" (UniqueName: \"kubernetes.io/projected/e775d3b4-8e2f-4255-8175-9664129b7998-kube-api-access-7jh9f\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.840795 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.841047 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.844089 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.853736 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jh9f\" (UniqueName: \"kubernetes.io/projected/e775d3b4-8e2f-4255-8175-9664129b7998-kube-api-access-7jh9f\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:24 crc kubenswrapper[4815]: I1205 09:55:24.984122 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:25 crc kubenswrapper[4815]: I1205 09:55:25.534939 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz"] Dec 05 09:55:25 crc kubenswrapper[4815]: I1205 09:55:25.591326 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" event={"ID":"e775d3b4-8e2f-4255-8175-9664129b7998","Type":"ContainerStarted","Data":"84dec56d00058cc8f02b837f005e1c6766140b019f7116b5a90454a3a9d75239"} Dec 05 09:55:25 crc kubenswrapper[4815]: I1205 09:55:25.593750 4815 generic.go:334] "Generic (PLEG): container finished" podID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" containerID="e04cd08e9eb6bc1dfd4063a5defcdeac81e3adbad58c07079ee0b9a6986426ce" exitCode=0 Dec 05 09:55:25 crc kubenswrapper[4815]: I1205 09:55:25.593801 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pbw8b" event={"ID":"d0b9fc78-4e68-43b4-83c8-9d1b703b834e","Type":"ContainerDied","Data":"e04cd08e9eb6bc1dfd4063a5defcdeac81e3adbad58c07079ee0b9a6986426ce"} Dec 05 09:55:26 crc kubenswrapper[4815]: I1205 09:55:26.604585 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" event={"ID":"e775d3b4-8e2f-4255-8175-9664129b7998","Type":"ContainerStarted","Data":"ea12a7ca2c2edf37c9879bde4bbbac3806296d3cd0020d9f70e3115dbf2a2a48"} Dec 05 09:55:26 crc kubenswrapper[4815]: I1205 09:55:26.612974 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pbw8b" event={"ID":"d0b9fc78-4e68-43b4-83c8-9d1b703b834e","Type":"ContainerStarted","Data":"12532b149458ec3f99551fa4fdb48915f7d527484b0ebebc22f762ca99009352"} Dec 05 09:55:26 crc kubenswrapper[4815]: I1205 09:55:26.633110 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" podStartSLOduration=1.990170563 podStartE2EDuration="2.633086527s" podCreationTimestamp="2025-12-05 09:55:24 +0000 UTC" firstStartedPulling="2025-12-05 09:55:25.538950068 +0000 UTC m=+2924.417556905" lastFinishedPulling="2025-12-05 09:55:26.181866032 +0000 UTC m=+2925.060472869" observedRunningTime="2025-12-05 09:55:26.62374129 +0000 UTC m=+2925.502348137" watchObservedRunningTime="2025-12-05 09:55:26.633086527 +0000 UTC m=+2925.511693374" Dec 05 09:55:26 crc kubenswrapper[4815]: I1205 09:55:26.649910 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pbw8b" podStartSLOduration=1.9237597819999999 podStartE2EDuration="4.64988718s" podCreationTimestamp="2025-12-05 09:55:22 +0000 UTC" firstStartedPulling="2025-12-05 09:55:23.572853099 +0000 UTC m=+2922.451459936" lastFinishedPulling="2025-12-05 09:55:26.298980497 +0000 UTC m=+2925.177587334" observedRunningTime="2025-12-05 09:55:26.642096445 +0000 UTC m=+2925.520703292" watchObservedRunningTime="2025-12-05 09:55:26.64988718 +0000 UTC m=+2925.528494017" Dec 05 09:55:32 crc kubenswrapper[4815]: I1205 09:55:32.704745 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:32 crc kubenswrapper[4815]: I1205 09:55:32.705344 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:32 crc kubenswrapper[4815]: I1205 09:55:32.765375 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:33 crc kubenswrapper[4815]: I1205 09:55:33.744148 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:33 crc kubenswrapper[4815]: I1205 09:55:33.811974 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pbw8b"] Dec 05 09:55:35 crc kubenswrapper[4815]: I1205 09:55:35.687574 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pbw8b" podUID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" containerName="registry-server" containerID="cri-o://12532b149458ec3f99551fa4fdb48915f7d527484b0ebebc22f762ca99009352" gracePeriod=2 Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.719223 4815 generic.go:334] "Generic (PLEG): container finished" podID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" containerID="12532b149458ec3f99551fa4fdb48915f7d527484b0ebebc22f762ca99009352" exitCode=0 Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.719757 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pbw8b" event={"ID":"d0b9fc78-4e68-43b4-83c8-9d1b703b834e","Type":"ContainerDied","Data":"12532b149458ec3f99551fa4fdb48915f7d527484b0ebebc22f762ca99009352"} Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.719784 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pbw8b" event={"ID":"d0b9fc78-4e68-43b4-83c8-9d1b703b834e","Type":"ContainerDied","Data":"ff2d287a7ac100b74506a2190f61eadd83cfc80dd7971e574db0bcf6b301cee5"} Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.719794 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff2d287a7ac100b74506a2190f61eadd83cfc80dd7971e574db0bcf6b301cee5" Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.775520 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.838116 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-268bv\" (UniqueName: \"kubernetes.io/projected/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-kube-api-access-268bv\") pod \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.838232 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-utilities\") pod \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.839045 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-utilities" (OuterVolumeSpecName: "utilities") pod "d0b9fc78-4e68-43b4-83c8-9d1b703b834e" (UID: "d0b9fc78-4e68-43b4-83c8-9d1b703b834e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.839137 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-catalog-content\") pod \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\" (UID: \"d0b9fc78-4e68-43b4-83c8-9d1b703b834e\") " Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.839670 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.859730 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-kube-api-access-268bv" (OuterVolumeSpecName: "kube-api-access-268bv") pod "d0b9fc78-4e68-43b4-83c8-9d1b703b834e" (UID: "d0b9fc78-4e68-43b4-83c8-9d1b703b834e"). InnerVolumeSpecName "kube-api-access-268bv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.896166 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0b9fc78-4e68-43b4-83c8-9d1b703b834e" (UID: "d0b9fc78-4e68-43b4-83c8-9d1b703b834e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.941193 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:36 crc kubenswrapper[4815]: I1205 09:55:36.941245 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-268bv\" (UniqueName: \"kubernetes.io/projected/d0b9fc78-4e68-43b4-83c8-9d1b703b834e-kube-api-access-268bv\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:37 crc kubenswrapper[4815]: I1205 09:55:37.740109 4815 generic.go:334] "Generic (PLEG): container finished" podID="e775d3b4-8e2f-4255-8175-9664129b7998" containerID="ea12a7ca2c2edf37c9879bde4bbbac3806296d3cd0020d9f70e3115dbf2a2a48" exitCode=0 Dec 05 09:55:37 crc kubenswrapper[4815]: I1205 09:55:37.740222 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" event={"ID":"e775d3b4-8e2f-4255-8175-9664129b7998","Type":"ContainerDied","Data":"ea12a7ca2c2edf37c9879bde4bbbac3806296d3cd0020d9f70e3115dbf2a2a48"} Dec 05 09:55:37 crc kubenswrapper[4815]: I1205 09:55:37.740652 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pbw8b" Dec 05 09:55:37 crc kubenswrapper[4815]: I1205 09:55:37.787243 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pbw8b"] Dec 05 09:55:37 crc kubenswrapper[4815]: I1205 09:55:37.794661 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pbw8b"] Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.214339 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.285726 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ceph\") pod \"e775d3b4-8e2f-4255-8175-9664129b7998\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.285839 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-inventory\") pod \"e775d3b4-8e2f-4255-8175-9664129b7998\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.285941 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ssh-key\") pod \"e775d3b4-8e2f-4255-8175-9664129b7998\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.286012 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jh9f\" (UniqueName: \"kubernetes.io/projected/e775d3b4-8e2f-4255-8175-9664129b7998-kube-api-access-7jh9f\") pod \"e775d3b4-8e2f-4255-8175-9664129b7998\" (UID: \"e775d3b4-8e2f-4255-8175-9664129b7998\") " Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.295152 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e775d3b4-8e2f-4255-8175-9664129b7998-kube-api-access-7jh9f" (OuterVolumeSpecName: "kube-api-access-7jh9f") pod "e775d3b4-8e2f-4255-8175-9664129b7998" (UID: "e775d3b4-8e2f-4255-8175-9664129b7998"). InnerVolumeSpecName "kube-api-access-7jh9f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.295643 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ceph" (OuterVolumeSpecName: "ceph") pod "e775d3b4-8e2f-4255-8175-9664129b7998" (UID: "e775d3b4-8e2f-4255-8175-9664129b7998"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.313524 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e775d3b4-8e2f-4255-8175-9664129b7998" (UID: "e775d3b4-8e2f-4255-8175-9664129b7998"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.319730 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-inventory" (OuterVolumeSpecName: "inventory") pod "e775d3b4-8e2f-4255-8175-9664129b7998" (UID: "e775d3b4-8e2f-4255-8175-9664129b7998"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.387452 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.387564 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jh9f\" (UniqueName: \"kubernetes.io/projected/e775d3b4-8e2f-4255-8175-9664129b7998-kube-api-access-7jh9f\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.387578 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.387586 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e775d3b4-8e2f-4255-8175-9664129b7998-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.431119 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" path="/var/lib/kubelet/pods/d0b9fc78-4e68-43b4-83c8-9d1b703b834e/volumes" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.763790 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" event={"ID":"e775d3b4-8e2f-4255-8175-9664129b7998","Type":"ContainerDied","Data":"84dec56d00058cc8f02b837f005e1c6766140b019f7116b5a90454a3a9d75239"} Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.764264 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84dec56d00058cc8f02b837f005e1c6766140b019f7116b5a90454a3a9d75239" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.763857 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.886839 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9"] Dec 05 09:55:39 crc kubenswrapper[4815]: E1205 09:55:39.887595 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" containerName="extract-content" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.887662 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" containerName="extract-content" Dec 05 09:55:39 crc kubenswrapper[4815]: E1205 09:55:39.887733 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" containerName="registry-server" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.887753 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" containerName="registry-server" Dec 05 09:55:39 crc kubenswrapper[4815]: E1205 09:55:39.887811 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" containerName="extract-utilities" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.887857 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" containerName="extract-utilities" Dec 05 09:55:39 crc kubenswrapper[4815]: E1205 09:55:39.887900 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e775d3b4-8e2f-4255-8175-9664129b7998" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.887920 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e775d3b4-8e2f-4255-8175-9664129b7998" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.888381 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e775d3b4-8e2f-4255-8175-9664129b7998" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.888423 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0b9fc78-4e68-43b4-83c8-9d1b703b834e" containerName="registry-server" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.889615 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.896102 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.896130 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.896188 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.896102 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.896416 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.896679 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.897063 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.898516 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.899341 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9"] Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.996822 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.996887 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.996913 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.996972 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.996991 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.997025 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.997118 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.997227 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.997289 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.997336 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hhd2\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-kube-api-access-2hhd2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.997432 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.997566 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:39 crc kubenswrapper[4815]: I1205 09:55:39.997615 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098567 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098615 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098637 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098657 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098705 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098736 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098762 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098788 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098906 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098932 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hhd2\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-kube-api-access-2hhd2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098962 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.098986 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.099007 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.103223 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.103330 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.103372 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.103694 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.105446 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.105836 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.107560 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.107799 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.108619 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.112980 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.113038 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.120275 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.122366 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hhd2\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-kube-api-access-2hhd2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.290416 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:55:40 crc kubenswrapper[4815]: I1205 09:55:40.816270 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9"] Dec 05 09:55:41 crc kubenswrapper[4815]: I1205 09:55:41.790737 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" event={"ID":"abe1e909-bfe1-4bab-9b1e-5ed3316288b6","Type":"ContainerStarted","Data":"d52c67f6b0aa4622d678fbce9eace11c360c6a3ebd51599b51a5f23d39f0b636"} Dec 05 09:55:41 crc kubenswrapper[4815]: I1205 09:55:41.791095 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" event={"ID":"abe1e909-bfe1-4bab-9b1e-5ed3316288b6","Type":"ContainerStarted","Data":"08f1511b027debf05ad94998d47dd1d791f0fa940bd41bdd4cf1c7f24542f437"} Dec 05 09:55:41 crc kubenswrapper[4815]: I1205 09:55:41.818205 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" podStartSLOduration=2.286262859 podStartE2EDuration="2.818173317s" podCreationTimestamp="2025-12-05 09:55:39 +0000 UTC" firstStartedPulling="2025-12-05 09:55:40.811950198 +0000 UTC m=+2939.690557035" lastFinishedPulling="2025-12-05 09:55:41.343860646 +0000 UTC m=+2940.222467493" observedRunningTime="2025-12-05 09:55:41.808007497 +0000 UTC m=+2940.686614354" watchObservedRunningTime="2025-12-05 09:55:41.818173317 +0000 UTC m=+2940.696780174" Dec 05 09:55:50 crc kubenswrapper[4815]: I1205 09:55:50.192513 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:55:50 crc kubenswrapper[4815]: I1205 09:55:50.193226 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:56:16 crc kubenswrapper[4815]: I1205 09:56:16.102433 4815 generic.go:334] "Generic (PLEG): container finished" podID="abe1e909-bfe1-4bab-9b1e-5ed3316288b6" containerID="d52c67f6b0aa4622d678fbce9eace11c360c6a3ebd51599b51a5f23d39f0b636" exitCode=0 Dec 05 09:56:16 crc kubenswrapper[4815]: I1205 09:56:16.102602 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" event={"ID":"abe1e909-bfe1-4bab-9b1e-5ed3316288b6","Type":"ContainerDied","Data":"d52c67f6b0aa4622d678fbce9eace11c360c6a3ebd51599b51a5f23d39f0b636"} Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.538515 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.675732 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-libvirt-combined-ca-bundle\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.675834 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-nova-combined-ca-bundle\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.675900 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ssh-key\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.675925 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-inventory\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.675981 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.676432 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-bootstrap-combined-ca-bundle\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.676538 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ovn-combined-ca-bundle\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.676574 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-neutron-metadata-combined-ca-bundle\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.676611 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.676643 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ceph\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.676694 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.676739 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2hhd2\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-kube-api-access-2hhd2\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.676796 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-repo-setup-combined-ca-bundle\") pod \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\" (UID: \"abe1e909-bfe1-4bab-9b1e-5ed3316288b6\") " Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.684259 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ceph" (OuterVolumeSpecName: "ceph") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.684358 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.684451 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.685126 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.686376 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.686407 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.686455 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.686981 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.687353 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.688378 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-kube-api-access-2hhd2" (OuterVolumeSpecName: "kube-api-access-2hhd2") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "kube-api-access-2hhd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.695657 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.715377 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-inventory" (OuterVolumeSpecName: "inventory") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.716699 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "abe1e909-bfe1-4bab-9b1e-5ed3316288b6" (UID: "abe1e909-bfe1-4bab-9b1e-5ed3316288b6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779449 4815 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779579 4815 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779597 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779610 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779621 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779633 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2hhd2\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-kube-api-access-2hhd2\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779641 4815 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779650 4815 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779658 4815 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779667 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779674 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779685 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:17 crc kubenswrapper[4815]: I1205 09:56:17.779695 4815 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/abe1e909-bfe1-4bab-9b1e-5ed3316288b6-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.157830 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" event={"ID":"abe1e909-bfe1-4bab-9b1e-5ed3316288b6","Type":"ContainerDied","Data":"08f1511b027debf05ad94998d47dd1d791f0fa940bd41bdd4cf1c7f24542f437"} Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.157876 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08f1511b027debf05ad94998d47dd1d791f0fa940bd41bdd4cf1c7f24542f437" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.157952 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.245150 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8"] Dec 05 09:56:18 crc kubenswrapper[4815]: E1205 09:56:18.245609 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abe1e909-bfe1-4bab-9b1e-5ed3316288b6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.245638 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="abe1e909-bfe1-4bab-9b1e-5ed3316288b6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.245871 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="abe1e909-bfe1-4bab-9b1e-5ed3316288b6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.246522 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.248768 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.249752 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.250291 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.250479 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.262931 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8"] Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.263623 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.391325 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.391379 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.391428 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.391878 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nxww\" (UniqueName: \"kubernetes.io/projected/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-kube-api-access-8nxww\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.494139 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.494220 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.494272 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.494341 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nxww\" (UniqueName: \"kubernetes.io/projected/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-kube-api-access-8nxww\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.499651 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.501159 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.506045 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.514478 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nxww\" (UniqueName: \"kubernetes.io/projected/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-kube-api-access-8nxww\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:18 crc kubenswrapper[4815]: I1205 09:56:18.564346 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:19 crc kubenswrapper[4815]: I1205 09:56:19.109855 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8"] Dec 05 09:56:19 crc kubenswrapper[4815]: I1205 09:56:19.170341 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" event={"ID":"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8","Type":"ContainerStarted","Data":"ab8887b81394ff1ed0f9358f8113d522f011ba2ea207cbe612e906d3fd88802a"} Dec 05 09:56:20 crc kubenswrapper[4815]: I1205 09:56:20.192289 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:56:20 crc kubenswrapper[4815]: I1205 09:56:20.193742 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:56:20 crc kubenswrapper[4815]: I1205 09:56:20.195291 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" event={"ID":"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8","Type":"ContainerStarted","Data":"fc8c433339187037cf4dcefe8ab37009b354464f3320242a2acc4048b8707596"} Dec 05 09:56:20 crc kubenswrapper[4815]: I1205 09:56:20.234259 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" podStartSLOduration=1.642673345 podStartE2EDuration="2.234238654s" podCreationTimestamp="2025-12-05 09:56:18 +0000 UTC" firstStartedPulling="2025-12-05 09:56:19.11631325 +0000 UTC m=+2977.994920087" lastFinishedPulling="2025-12-05 09:56:19.707878549 +0000 UTC m=+2978.586485396" observedRunningTime="2025-12-05 09:56:20.220369712 +0000 UTC m=+2979.098976549" watchObservedRunningTime="2025-12-05 09:56:20.234238654 +0000 UTC m=+2979.112845491" Dec 05 09:56:25 crc kubenswrapper[4815]: I1205 09:56:25.907468 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" podUID="b7558b70-107c-48ea-ac22-e42b1bcdf47d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.84:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:56:25 crc kubenswrapper[4815]: I1205 09:56:25.911155 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-55c85496f5b4xkh" podUID="b7558b70-107c-48ea-ac22-e42b1bcdf47d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.84:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 09:56:27 crc kubenswrapper[4815]: I1205 09:56:27.924046 4815 generic.go:334] "Generic (PLEG): container finished" podID="dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8" containerID="fc8c433339187037cf4dcefe8ab37009b354464f3320242a2acc4048b8707596" exitCode=0 Dec 05 09:56:27 crc kubenswrapper[4815]: I1205 09:56:27.924132 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" event={"ID":"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8","Type":"ContainerDied","Data":"fc8c433339187037cf4dcefe8ab37009b354464f3320242a2acc4048b8707596"} Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.349384 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.454069 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ceph\") pod \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.454225 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ssh-key\") pod \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.454281 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-inventory\") pod \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.454330 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nxww\" (UniqueName: \"kubernetes.io/projected/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-kube-api-access-8nxww\") pod \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\" (UID: \"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8\") " Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.459652 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-kube-api-access-8nxww" (OuterVolumeSpecName: "kube-api-access-8nxww") pod "dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8" (UID: "dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8"). InnerVolumeSpecName "kube-api-access-8nxww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.476254 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ceph" (OuterVolumeSpecName: "ceph") pod "dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8" (UID: "dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.479409 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8" (UID: "dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.484254 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-inventory" (OuterVolumeSpecName: "inventory") pod "dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8" (UID: "dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.559553 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.559587 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.559597 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nxww\" (UniqueName: \"kubernetes.io/projected/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-kube-api-access-8nxww\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.559606 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.944331 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" event={"ID":"dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8","Type":"ContainerDied","Data":"ab8887b81394ff1ed0f9358f8113d522f011ba2ea207cbe612e906d3fd88802a"} Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.944375 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab8887b81394ff1ed0f9358f8113d522f011ba2ea207cbe612e906d3fd88802a" Dec 05 09:56:29 crc kubenswrapper[4815]: I1205 09:56:29.944381 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.043329 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr"] Dec 05 09:56:30 crc kubenswrapper[4815]: E1205 09:56:30.043717 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.043736 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.043975 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.044606 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.048572 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.048920 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.049090 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.049102 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.049259 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.049986 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.066865 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr"] Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.174992 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.175183 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.175314 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.175353 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.175420 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.175436 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sxs8\" (UniqueName: \"kubernetes.io/projected/b3cdccbb-a413-4128-8d79-34ac79a10a45-kube-api-access-6sxs8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.276841 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.276885 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.276922 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sxs8\" (UniqueName: \"kubernetes.io/projected/b3cdccbb-a413-4128-8d79-34ac79a10a45-kube-api-access-6sxs8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.276941 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.276992 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.277063 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.278687 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.281986 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.282554 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.292104 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.292925 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.294981 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sxs8\" (UniqueName: \"kubernetes.io/projected/b3cdccbb-a413-4128-8d79-34ac79a10a45-kube-api-access-6sxs8\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-kv4qr\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.364126 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:56:30 crc kubenswrapper[4815]: I1205 09:56:30.954069 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr"] Dec 05 09:56:31 crc kubenswrapper[4815]: I1205 09:56:31.961054 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" event={"ID":"b3cdccbb-a413-4128-8d79-34ac79a10a45","Type":"ContainerStarted","Data":"fcb6c287cbad1a0e53f5a22fa422c13b84fe3a66c3d358f40e4d06691f78fb80"} Dec 05 09:56:32 crc kubenswrapper[4815]: I1205 09:56:32.984409 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" event={"ID":"b3cdccbb-a413-4128-8d79-34ac79a10a45","Type":"ContainerStarted","Data":"40c92bac5910fbeb2ab71e6b853bfeeb80d34e66f042756efb18128540e9bc30"} Dec 05 09:56:33 crc kubenswrapper[4815]: I1205 09:56:33.007138 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" podStartSLOduration=1.782030803 podStartE2EDuration="3.007117218s" podCreationTimestamp="2025-12-05 09:56:30 +0000 UTC" firstStartedPulling="2025-12-05 09:56:30.961174209 +0000 UTC m=+2989.839781036" lastFinishedPulling="2025-12-05 09:56:32.186260614 +0000 UTC m=+2991.064867451" observedRunningTime="2025-12-05 09:56:33.005289148 +0000 UTC m=+2991.883895995" watchObservedRunningTime="2025-12-05 09:56:33.007117218 +0000 UTC m=+2991.885724055" Dec 05 09:56:50 crc kubenswrapper[4815]: I1205 09:56:50.192308 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 09:56:50 crc kubenswrapper[4815]: I1205 09:56:50.192849 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 09:56:50 crc kubenswrapper[4815]: I1205 09:56:50.192905 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 09:56:50 crc kubenswrapper[4815]: I1205 09:56:50.193731 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 09:56:50 crc kubenswrapper[4815]: I1205 09:56:50.193793 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" gracePeriod=600 Dec 05 09:56:50 crc kubenswrapper[4815]: E1205 09:56:50.571869 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:56:51 crc kubenswrapper[4815]: I1205 09:56:51.172792 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" exitCode=0 Dec 05 09:56:51 crc kubenswrapper[4815]: I1205 09:56:51.172839 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754"} Dec 05 09:56:51 crc kubenswrapper[4815]: I1205 09:56:51.172903 4815 scope.go:117] "RemoveContainer" containerID="92d524fe5b80f191d3ca613cde5455e954b4de4abdbc8a66cdde8df7e0934c6c" Dec 05 09:56:51 crc kubenswrapper[4815]: I1205 09:56:51.173774 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:56:51 crc kubenswrapper[4815]: E1205 09:56:51.174162 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:57:05 crc kubenswrapper[4815]: I1205 09:57:05.418360 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:57:05 crc kubenswrapper[4815]: E1205 09:57:05.419134 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:57:19 crc kubenswrapper[4815]: I1205 09:57:19.418719 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:57:19 crc kubenswrapper[4815]: E1205 09:57:19.419444 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:57:31 crc kubenswrapper[4815]: I1205 09:57:31.425227 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:57:31 crc kubenswrapper[4815]: E1205 09:57:31.426046 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.347314 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mpxtd"] Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.350021 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.356965 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mpxtd"] Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.517550 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-utilities\") pod \"redhat-operators-mpxtd\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.517875 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-catalog-content\") pod \"redhat-operators-mpxtd\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.518052 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bv9p\" (UniqueName: \"kubernetes.io/projected/1b358236-5d59-48f1-b4f3-25f37416b5de-kube-api-access-9bv9p\") pod \"redhat-operators-mpxtd\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.619246 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-catalog-content\") pod \"redhat-operators-mpxtd\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.619599 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bv9p\" (UniqueName: \"kubernetes.io/projected/1b358236-5d59-48f1-b4f3-25f37416b5de-kube-api-access-9bv9p\") pod \"redhat-operators-mpxtd\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.619753 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-utilities\") pod \"redhat-operators-mpxtd\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.620335 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-utilities\") pod \"redhat-operators-mpxtd\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.620714 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-catalog-content\") pod \"redhat-operators-mpxtd\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.644088 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bv9p\" (UniqueName: \"kubernetes.io/projected/1b358236-5d59-48f1-b4f3-25f37416b5de-kube-api-access-9bv9p\") pod \"redhat-operators-mpxtd\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:40 crc kubenswrapper[4815]: I1205 09:57:40.673919 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:41 crc kubenswrapper[4815]: I1205 09:57:41.156035 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mpxtd"] Dec 05 09:57:41 crc kubenswrapper[4815]: I1205 09:57:41.714666 4815 generic.go:334] "Generic (PLEG): container finished" podID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerID="369859423f3b0f62c8aff2851b04ea52c46b67460007a4087fb14dc14c5ad730" exitCode=0 Dec 05 09:57:41 crc kubenswrapper[4815]: I1205 09:57:41.714761 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpxtd" event={"ID":"1b358236-5d59-48f1-b4f3-25f37416b5de","Type":"ContainerDied","Data":"369859423f3b0f62c8aff2851b04ea52c46b67460007a4087fb14dc14c5ad730"} Dec 05 09:57:41 crc kubenswrapper[4815]: I1205 09:57:41.715629 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpxtd" event={"ID":"1b358236-5d59-48f1-b4f3-25f37416b5de","Type":"ContainerStarted","Data":"aa6e95576d6926c6364aba497d3a4cfc24c4bf86e863eccaf89b8b030fbc2e1f"} Dec 05 09:57:42 crc kubenswrapper[4815]: I1205 09:57:42.728204 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpxtd" event={"ID":"1b358236-5d59-48f1-b4f3-25f37416b5de","Type":"ContainerStarted","Data":"76913a66cc34c4cf00676d306ca50fc058ccf0487743c39ebf5824ae8a2c2ea7"} Dec 05 09:57:46 crc kubenswrapper[4815]: I1205 09:57:46.445805 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:57:46 crc kubenswrapper[4815]: E1205 09:57:46.446821 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:57:47 crc kubenswrapper[4815]: I1205 09:57:47.776324 4815 generic.go:334] "Generic (PLEG): container finished" podID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerID="76913a66cc34c4cf00676d306ca50fc058ccf0487743c39ebf5824ae8a2c2ea7" exitCode=0 Dec 05 09:57:47 crc kubenswrapper[4815]: I1205 09:57:47.776404 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpxtd" event={"ID":"1b358236-5d59-48f1-b4f3-25f37416b5de","Type":"ContainerDied","Data":"76913a66cc34c4cf00676d306ca50fc058ccf0487743c39ebf5824ae8a2c2ea7"} Dec 05 09:57:48 crc kubenswrapper[4815]: I1205 09:57:48.796896 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpxtd" event={"ID":"1b358236-5d59-48f1-b4f3-25f37416b5de","Type":"ContainerStarted","Data":"52d30b01bf9604ca1f1fe0dbbe17bf8809878930b938c84aa839fe13a8464368"} Dec 05 09:57:48 crc kubenswrapper[4815]: I1205 09:57:48.823185 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mpxtd" podStartSLOduration=2.29653747 podStartE2EDuration="8.82315614s" podCreationTimestamp="2025-12-05 09:57:40 +0000 UTC" firstStartedPulling="2025-12-05 09:57:41.717249484 +0000 UTC m=+3060.595856321" lastFinishedPulling="2025-12-05 09:57:48.243868154 +0000 UTC m=+3067.122474991" observedRunningTime="2025-12-05 09:57:48.814566334 +0000 UTC m=+3067.693173191" watchObservedRunningTime="2025-12-05 09:57:48.82315614 +0000 UTC m=+3067.701762977" Dec 05 09:57:50 crc kubenswrapper[4815]: I1205 09:57:50.676689 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:50 crc kubenswrapper[4815]: I1205 09:57:50.676967 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:57:51 crc kubenswrapper[4815]: I1205 09:57:51.723093 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mpxtd" podUID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerName="registry-server" probeResult="failure" output=< Dec 05 09:57:51 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 09:57:51 crc kubenswrapper[4815]: > Dec 05 09:57:56 crc kubenswrapper[4815]: I1205 09:57:56.895863 4815 generic.go:334] "Generic (PLEG): container finished" podID="b3cdccbb-a413-4128-8d79-34ac79a10a45" containerID="40c92bac5910fbeb2ab71e6b853bfeeb80d34e66f042756efb18128540e9bc30" exitCode=0 Dec 05 09:57:56 crc kubenswrapper[4815]: I1205 09:57:56.895934 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" event={"ID":"b3cdccbb-a413-4128-8d79-34ac79a10a45","Type":"ContainerDied","Data":"40c92bac5910fbeb2ab71e6b853bfeeb80d34e66f042756efb18128540e9bc30"} Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.269671 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.419465 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:57:58 crc kubenswrapper[4815]: E1205 09:57:58.419932 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.434392 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-inventory\") pod \"b3cdccbb-a413-4128-8d79-34ac79a10a45\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.434435 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovncontroller-config-0\") pod \"b3cdccbb-a413-4128-8d79-34ac79a10a45\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.434519 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ssh-key\") pod \"b3cdccbb-a413-4128-8d79-34ac79a10a45\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.434624 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovn-combined-ca-bundle\") pod \"b3cdccbb-a413-4128-8d79-34ac79a10a45\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.434664 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6sxs8\" (UniqueName: \"kubernetes.io/projected/b3cdccbb-a413-4128-8d79-34ac79a10a45-kube-api-access-6sxs8\") pod \"b3cdccbb-a413-4128-8d79-34ac79a10a45\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.434786 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ceph\") pod \"b3cdccbb-a413-4128-8d79-34ac79a10a45\" (UID: \"b3cdccbb-a413-4128-8d79-34ac79a10a45\") " Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.440214 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ceph" (OuterVolumeSpecName: "ceph") pod "b3cdccbb-a413-4128-8d79-34ac79a10a45" (UID: "b3cdccbb-a413-4128-8d79-34ac79a10a45"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.440533 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "b3cdccbb-a413-4128-8d79-34ac79a10a45" (UID: "b3cdccbb-a413-4128-8d79-34ac79a10a45"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.441129 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3cdccbb-a413-4128-8d79-34ac79a10a45-kube-api-access-6sxs8" (OuterVolumeSpecName: "kube-api-access-6sxs8") pod "b3cdccbb-a413-4128-8d79-34ac79a10a45" (UID: "b3cdccbb-a413-4128-8d79-34ac79a10a45"). InnerVolumeSpecName "kube-api-access-6sxs8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.461550 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "b3cdccbb-a413-4128-8d79-34ac79a10a45" (UID: "b3cdccbb-a413-4128-8d79-34ac79a10a45"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.464691 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b3cdccbb-a413-4128-8d79-34ac79a10a45" (UID: "b3cdccbb-a413-4128-8d79-34ac79a10a45"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.465862 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-inventory" (OuterVolumeSpecName: "inventory") pod "b3cdccbb-a413-4128-8d79-34ac79a10a45" (UID: "b3cdccbb-a413-4128-8d79-34ac79a10a45"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.537858 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.538209 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.538260 4815 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.538276 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.538306 4815 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b3cdccbb-a413-4128-8d79-34ac79a10a45-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.538320 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6sxs8\" (UniqueName: \"kubernetes.io/projected/b3cdccbb-a413-4128-8d79-34ac79a10a45-kube-api-access-6sxs8\") on node \"crc\" DevicePath \"\"" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.914087 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" event={"ID":"b3cdccbb-a413-4128-8d79-34ac79a10a45","Type":"ContainerDied","Data":"fcb6c287cbad1a0e53f5a22fa422c13b84fe3a66c3d358f40e4d06691f78fb80"} Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.914136 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fcb6c287cbad1a0e53f5a22fa422c13b84fe3a66c3d358f40e4d06691f78fb80" Dec 05 09:57:58 crc kubenswrapper[4815]: I1205 09:57:58.914228 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-kv4qr" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.148328 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n"] Dec 05 09:57:59 crc kubenswrapper[4815]: E1205 09:57:59.148918 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3cdccbb-a413-4128-8d79-34ac79a10a45" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.148944 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3cdccbb-a413-4128-8d79-34ac79a10a45" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.149131 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3cdccbb-a413-4128-8d79-34ac79a10a45" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.149798 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.153707 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.153807 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.153842 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.153909 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.153706 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.154336 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.155724 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.177191 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n"] Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.352476 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.352664 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.352796 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.352831 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.352870 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hp4g5\" (UniqueName: \"kubernetes.io/projected/76992009-c2d1-4e3b-be35-70fc39e9f999-kube-api-access-hp4g5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.353016 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.353324 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.456618 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.456747 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.456889 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.457043 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.457125 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.457184 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hp4g5\" (UniqueName: \"kubernetes.io/projected/76992009-c2d1-4e3b-be35-70fc39e9f999-kube-api-access-hp4g5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.457269 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.462553 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.463142 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.463789 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.462665 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.472239 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.472878 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.483260 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hp4g5\" (UniqueName: \"kubernetes.io/projected/76992009-c2d1-4e3b-be35-70fc39e9f999-kube-api-access-hp4g5\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:57:59 crc kubenswrapper[4815]: I1205 09:57:59.764658 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:58:00 crc kubenswrapper[4815]: I1205 09:58:00.358087 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n"] Dec 05 09:58:00 crc kubenswrapper[4815]: I1205 09:58:00.744874 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:58:00 crc kubenswrapper[4815]: I1205 09:58:00.798673 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:58:00 crc kubenswrapper[4815]: I1205 09:58:00.930808 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" event={"ID":"76992009-c2d1-4e3b-be35-70fc39e9f999","Type":"ContainerStarted","Data":"67ce6711718fe40ad11fc0ae7a0c3333012bd75d66262a1f51a33367063d9ec4"} Dec 05 09:58:00 crc kubenswrapper[4815]: I1205 09:58:00.985785 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mpxtd"] Dec 05 09:58:01 crc kubenswrapper[4815]: I1205 09:58:01.941043 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" event={"ID":"76992009-c2d1-4e3b-be35-70fc39e9f999","Type":"ContainerStarted","Data":"3b9545109e41333c8fb2fadd7f6ce04c92d11dab4ab201e31eb9bbacd4037beb"} Dec 05 09:58:01 crc kubenswrapper[4815]: I1205 09:58:01.941552 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mpxtd" podUID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerName="registry-server" containerID="cri-o://52d30b01bf9604ca1f1fe0dbbe17bf8809878930b938c84aa839fe13a8464368" gracePeriod=2 Dec 05 09:58:01 crc kubenswrapper[4815]: I1205 09:58:01.972329 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" podStartSLOduration=2.080701507 podStartE2EDuration="2.972280028s" podCreationTimestamp="2025-12-05 09:57:59 +0000 UTC" firstStartedPulling="2025-12-05 09:58:00.364687202 +0000 UTC m=+3079.243294059" lastFinishedPulling="2025-12-05 09:58:01.256265743 +0000 UTC m=+3080.134872580" observedRunningTime="2025-12-05 09:58:01.965942064 +0000 UTC m=+3080.844548911" watchObservedRunningTime="2025-12-05 09:58:01.972280028 +0000 UTC m=+3080.850886865" Dec 05 09:58:02 crc kubenswrapper[4815]: I1205 09:58:02.984612 4815 generic.go:334] "Generic (PLEG): container finished" podID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerID="52d30b01bf9604ca1f1fe0dbbe17bf8809878930b938c84aa839fe13a8464368" exitCode=0 Dec 05 09:58:02 crc kubenswrapper[4815]: I1205 09:58:02.984823 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpxtd" event={"ID":"1b358236-5d59-48f1-b4f3-25f37416b5de","Type":"ContainerDied","Data":"52d30b01bf9604ca1f1fe0dbbe17bf8809878930b938c84aa839fe13a8464368"} Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.219204 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.330769 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-utilities\") pod \"1b358236-5d59-48f1-b4f3-25f37416b5de\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.331137 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-catalog-content\") pod \"1b358236-5d59-48f1-b4f3-25f37416b5de\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.331402 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9bv9p\" (UniqueName: \"kubernetes.io/projected/1b358236-5d59-48f1-b4f3-25f37416b5de-kube-api-access-9bv9p\") pod \"1b358236-5d59-48f1-b4f3-25f37416b5de\" (UID: \"1b358236-5d59-48f1-b4f3-25f37416b5de\") " Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.333818 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-utilities" (OuterVolumeSpecName: "utilities") pod "1b358236-5d59-48f1-b4f3-25f37416b5de" (UID: "1b358236-5d59-48f1-b4f3-25f37416b5de"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.372446 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b358236-5d59-48f1-b4f3-25f37416b5de-kube-api-access-9bv9p" (OuterVolumeSpecName: "kube-api-access-9bv9p") pod "1b358236-5d59-48f1-b4f3-25f37416b5de" (UID: "1b358236-5d59-48f1-b4f3-25f37416b5de"). InnerVolumeSpecName "kube-api-access-9bv9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.435959 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9bv9p\" (UniqueName: \"kubernetes.io/projected/1b358236-5d59-48f1-b4f3-25f37416b5de-kube-api-access-9bv9p\") on node \"crc\" DevicePath \"\"" Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.436028 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.495032 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b358236-5d59-48f1-b4f3-25f37416b5de" (UID: "1b358236-5d59-48f1-b4f3-25f37416b5de"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.537674 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b358236-5d59-48f1-b4f3-25f37416b5de-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.994931 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mpxtd" event={"ID":"1b358236-5d59-48f1-b4f3-25f37416b5de","Type":"ContainerDied","Data":"aa6e95576d6926c6364aba497d3a4cfc24c4bf86e863eccaf89b8b030fbc2e1f"} Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.994986 4815 scope.go:117] "RemoveContainer" containerID="52d30b01bf9604ca1f1fe0dbbe17bf8809878930b938c84aa839fe13a8464368" Dec 05 09:58:03 crc kubenswrapper[4815]: I1205 09:58:03.994988 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mpxtd" Dec 05 09:58:04 crc kubenswrapper[4815]: I1205 09:58:04.031793 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mpxtd"] Dec 05 09:58:04 crc kubenswrapper[4815]: I1205 09:58:04.031898 4815 scope.go:117] "RemoveContainer" containerID="76913a66cc34c4cf00676d306ca50fc058ccf0487743c39ebf5824ae8a2c2ea7" Dec 05 09:58:04 crc kubenswrapper[4815]: I1205 09:58:04.052114 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mpxtd"] Dec 05 09:58:04 crc kubenswrapper[4815]: I1205 09:58:04.070618 4815 scope.go:117] "RemoveContainer" containerID="369859423f3b0f62c8aff2851b04ea52c46b67460007a4087fb14dc14c5ad730" Dec 05 09:58:05 crc kubenswrapper[4815]: I1205 09:58:05.431840 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b358236-5d59-48f1-b4f3-25f37416b5de" path="/var/lib/kubelet/pods/1b358236-5d59-48f1-b4f3-25f37416b5de/volumes" Dec 05 09:58:09 crc kubenswrapper[4815]: I1205 09:58:09.418542 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:58:09 crc kubenswrapper[4815]: E1205 09:58:09.419334 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:58:21 crc kubenswrapper[4815]: I1205 09:58:21.424080 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:58:21 crc kubenswrapper[4815]: E1205 09:58:21.424812 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:58:32 crc kubenswrapper[4815]: I1205 09:58:32.419124 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:58:32 crc kubenswrapper[4815]: E1205 09:58:32.419885 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:58:45 crc kubenswrapper[4815]: I1205 09:58:45.419423 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:58:45 crc kubenswrapper[4815]: E1205 09:58:45.420340 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:58:56 crc kubenswrapper[4815]: I1205 09:58:56.418872 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:58:56 crc kubenswrapper[4815]: E1205 09:58:56.420454 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:59:10 crc kubenswrapper[4815]: I1205 09:59:10.418594 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:59:10 crc kubenswrapper[4815]: E1205 09:59:10.419308 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:59:19 crc kubenswrapper[4815]: I1205 09:59:19.658228 4815 generic.go:334] "Generic (PLEG): container finished" podID="76992009-c2d1-4e3b-be35-70fc39e9f999" containerID="3b9545109e41333c8fb2fadd7f6ce04c92d11dab4ab201e31eb9bbacd4037beb" exitCode=0 Dec 05 09:59:19 crc kubenswrapper[4815]: I1205 09:59:19.658289 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" event={"ID":"76992009-c2d1-4e3b-be35-70fc39e9f999","Type":"ContainerDied","Data":"3b9545109e41333c8fb2fadd7f6ce04c92d11dab4ab201e31eb9bbacd4037beb"} Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.150808 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.338282 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hp4g5\" (UniqueName: \"kubernetes.io/projected/76992009-c2d1-4e3b-be35-70fc39e9f999-kube-api-access-hp4g5\") pod \"76992009-c2d1-4e3b-be35-70fc39e9f999\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.338443 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-ovn-metadata-agent-neutron-config-0\") pod \"76992009-c2d1-4e3b-be35-70fc39e9f999\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.338484 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-inventory\") pod \"76992009-c2d1-4e3b-be35-70fc39e9f999\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.338521 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-nova-metadata-neutron-config-0\") pod \"76992009-c2d1-4e3b-be35-70fc39e9f999\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.338554 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-metadata-combined-ca-bundle\") pod \"76992009-c2d1-4e3b-be35-70fc39e9f999\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.338578 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ssh-key\") pod \"76992009-c2d1-4e3b-be35-70fc39e9f999\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.338645 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ceph\") pod \"76992009-c2d1-4e3b-be35-70fc39e9f999\" (UID: \"76992009-c2d1-4e3b-be35-70fc39e9f999\") " Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.345336 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "76992009-c2d1-4e3b-be35-70fc39e9f999" (UID: "76992009-c2d1-4e3b-be35-70fc39e9f999"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.346411 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ceph" (OuterVolumeSpecName: "ceph") pod "76992009-c2d1-4e3b-be35-70fc39e9f999" (UID: "76992009-c2d1-4e3b-be35-70fc39e9f999"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.356578 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76992009-c2d1-4e3b-be35-70fc39e9f999-kube-api-access-hp4g5" (OuterVolumeSpecName: "kube-api-access-hp4g5") pod "76992009-c2d1-4e3b-be35-70fc39e9f999" (UID: "76992009-c2d1-4e3b-be35-70fc39e9f999"). InnerVolumeSpecName "kube-api-access-hp4g5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.369406 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "76992009-c2d1-4e3b-be35-70fc39e9f999" (UID: "76992009-c2d1-4e3b-be35-70fc39e9f999"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.370834 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "76992009-c2d1-4e3b-be35-70fc39e9f999" (UID: "76992009-c2d1-4e3b-be35-70fc39e9f999"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.371725 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-inventory" (OuterVolumeSpecName: "inventory") pod "76992009-c2d1-4e3b-be35-70fc39e9f999" (UID: "76992009-c2d1-4e3b-be35-70fc39e9f999"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.375515 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "76992009-c2d1-4e3b-be35-70fc39e9f999" (UID: "76992009-c2d1-4e3b-be35-70fc39e9f999"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.429947 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:59:21 crc kubenswrapper[4815]: E1205 09:59:21.430253 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.442355 4815 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.442405 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.442420 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.442431 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hp4g5\" (UniqueName: \"kubernetes.io/projected/76992009-c2d1-4e3b-be35-70fc39e9f999-kube-api-access-hp4g5\") on node \"crc\" DevicePath \"\"" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.442443 4815 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.442455 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.442467 4815 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/76992009-c2d1-4e3b-be35-70fc39e9f999-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.679311 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" event={"ID":"76992009-c2d1-4e3b-be35-70fc39e9f999","Type":"ContainerDied","Data":"67ce6711718fe40ad11fc0ae7a0c3333012bd75d66262a1f51a33367063d9ec4"} Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.679360 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="67ce6711718fe40ad11fc0ae7a0c3333012bd75d66262a1f51a33367063d9ec4" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.679429 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.794788 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs"] Dec 05 09:59:21 crc kubenswrapper[4815]: E1205 09:59:21.795662 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerName="extract-utilities" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.795798 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerName="extract-utilities" Dec 05 09:59:21 crc kubenswrapper[4815]: E1205 09:59:21.795909 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76992009-c2d1-4e3b-be35-70fc39e9f999" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.796002 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="76992009-c2d1-4e3b-be35-70fc39e9f999" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 05 09:59:21 crc kubenswrapper[4815]: E1205 09:59:21.796095 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerName="extract-content" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.796178 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerName="extract-content" Dec 05 09:59:21 crc kubenswrapper[4815]: E1205 09:59:21.796268 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerName="registry-server" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.796352 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerName="registry-server" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.796677 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b358236-5d59-48f1-b4f3-25f37416b5de" containerName="registry-server" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.796791 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="76992009-c2d1-4e3b-be35-70fc39e9f999" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.797520 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.804126 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.804156 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.804304 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.804392 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.804676 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.804757 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.815319 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs"] Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.954013 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.954129 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.954235 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc9l6\" (UniqueName: \"kubernetes.io/projected/72f92018-4b16-474a-b5de-8fd124dd857b-kube-api-access-xc9l6\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.954264 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.954290 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:21 crc kubenswrapper[4815]: I1205 09:59:21.954367 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.056615 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.056731 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.056807 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.056913 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc9l6\" (UniqueName: \"kubernetes.io/projected/72f92018-4b16-474a-b5de-8fd124dd857b-kube-api-access-xc9l6\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.056940 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.056968 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.062559 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.062713 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.072377 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.072820 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.073338 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.076610 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc9l6\" (UniqueName: \"kubernetes.io/projected/72f92018-4b16-474a-b5de-8fd124dd857b-kube-api-access-xc9l6\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-kknjs\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.125939 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.642819 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs"] Dec 05 09:59:22 crc kubenswrapper[4815]: I1205 09:59:22.688817 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" event={"ID":"72f92018-4b16-474a-b5de-8fd124dd857b","Type":"ContainerStarted","Data":"feb4a11b92871d9ec65088847a7a5738b2cb1046682cff42b5fbfa105ce57f7c"} Dec 05 09:59:27 crc kubenswrapper[4815]: I1205 09:59:27.732416 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" event={"ID":"72f92018-4b16-474a-b5de-8fd124dd857b","Type":"ContainerStarted","Data":"aaa542e1ea30ca7bf409a862fb65a0a5c46f6f53c202c0a75445e1d7c46d1498"} Dec 05 09:59:27 crc kubenswrapper[4815]: I1205 09:59:27.765989 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" podStartSLOduration=2.943338357 podStartE2EDuration="6.765966767s" podCreationTimestamp="2025-12-05 09:59:21 +0000 UTC" firstStartedPulling="2025-12-05 09:59:22.651681936 +0000 UTC m=+3161.530288773" lastFinishedPulling="2025-12-05 09:59:26.474310356 +0000 UTC m=+3165.352917183" observedRunningTime="2025-12-05 09:59:27.756543468 +0000 UTC m=+3166.635150305" watchObservedRunningTime="2025-12-05 09:59:27.765966767 +0000 UTC m=+3166.644573604" Dec 05 09:59:32 crc kubenswrapper[4815]: I1205 09:59:32.418678 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:59:32 crc kubenswrapper[4815]: E1205 09:59:32.419222 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:59:45 crc kubenswrapper[4815]: I1205 09:59:45.418547 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:59:45 crc kubenswrapper[4815]: E1205 09:59:45.419394 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 09:59:56 crc kubenswrapper[4815]: I1205 09:59:56.418866 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 09:59:56 crc kubenswrapper[4815]: E1205 09:59:56.419801 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.153831 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb"] Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.155222 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.157874 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.158395 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.170125 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb"] Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.306551 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e42a353-a6f5-427a-bff2-64f6a9b80660-config-volume\") pod \"collect-profiles-29415480-w77cb\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.307014 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e42a353-a6f5-427a-bff2-64f6a9b80660-secret-volume\") pod \"collect-profiles-29415480-w77cb\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.307055 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s5hd\" (UniqueName: \"kubernetes.io/projected/1e42a353-a6f5-427a-bff2-64f6a9b80660-kube-api-access-5s5hd\") pod \"collect-profiles-29415480-w77cb\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.408307 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e42a353-a6f5-427a-bff2-64f6a9b80660-secret-volume\") pod \"collect-profiles-29415480-w77cb\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.408369 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s5hd\" (UniqueName: \"kubernetes.io/projected/1e42a353-a6f5-427a-bff2-64f6a9b80660-kube-api-access-5s5hd\") pod \"collect-profiles-29415480-w77cb\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.408457 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e42a353-a6f5-427a-bff2-64f6a9b80660-config-volume\") pod \"collect-profiles-29415480-w77cb\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.409595 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e42a353-a6f5-427a-bff2-64f6a9b80660-config-volume\") pod \"collect-profiles-29415480-w77cb\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.416352 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e42a353-a6f5-427a-bff2-64f6a9b80660-secret-volume\") pod \"collect-profiles-29415480-w77cb\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.430207 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s5hd\" (UniqueName: \"kubernetes.io/projected/1e42a353-a6f5-427a-bff2-64f6a9b80660-kube-api-access-5s5hd\") pod \"collect-profiles-29415480-w77cb\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.480713 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:00 crc kubenswrapper[4815]: I1205 10:00:00.933239 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb"] Dec 05 10:00:01 crc kubenswrapper[4815]: I1205 10:00:01.064010 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" event={"ID":"1e42a353-a6f5-427a-bff2-64f6a9b80660","Type":"ContainerStarted","Data":"71f6d0504c167ddf5850aa1e1ebf1721e5b9a5c5f692f5cac9593d9bbdc6dbad"} Dec 05 10:00:02 crc kubenswrapper[4815]: I1205 10:00:02.079146 4815 generic.go:334] "Generic (PLEG): container finished" podID="1e42a353-a6f5-427a-bff2-64f6a9b80660" containerID="57e2fa5a7af642c7f38603e6e3672abf5849358e0818b54d1851adae717aabfb" exitCode=0 Dec 05 10:00:02 crc kubenswrapper[4815]: I1205 10:00:02.079218 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" event={"ID":"1e42a353-a6f5-427a-bff2-64f6a9b80660","Type":"ContainerDied","Data":"57e2fa5a7af642c7f38603e6e3672abf5849358e0818b54d1851adae717aabfb"} Dec 05 10:00:03 crc kubenswrapper[4815]: I1205 10:00:03.465812 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:03 crc kubenswrapper[4815]: I1205 10:00:03.583954 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e42a353-a6f5-427a-bff2-64f6a9b80660-secret-volume\") pod \"1e42a353-a6f5-427a-bff2-64f6a9b80660\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " Dec 05 10:00:03 crc kubenswrapper[4815]: I1205 10:00:03.584068 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e42a353-a6f5-427a-bff2-64f6a9b80660-config-volume\") pod \"1e42a353-a6f5-427a-bff2-64f6a9b80660\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " Dec 05 10:00:03 crc kubenswrapper[4815]: I1205 10:00:03.584104 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5s5hd\" (UniqueName: \"kubernetes.io/projected/1e42a353-a6f5-427a-bff2-64f6a9b80660-kube-api-access-5s5hd\") pod \"1e42a353-a6f5-427a-bff2-64f6a9b80660\" (UID: \"1e42a353-a6f5-427a-bff2-64f6a9b80660\") " Dec 05 10:00:03 crc kubenswrapper[4815]: I1205 10:00:03.585484 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e42a353-a6f5-427a-bff2-64f6a9b80660-config-volume" (OuterVolumeSpecName: "config-volume") pod "1e42a353-a6f5-427a-bff2-64f6a9b80660" (UID: "1e42a353-a6f5-427a-bff2-64f6a9b80660"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:00:03 crc kubenswrapper[4815]: I1205 10:00:03.589929 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1e42a353-a6f5-427a-bff2-64f6a9b80660-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1e42a353-a6f5-427a-bff2-64f6a9b80660" (UID: "1e42a353-a6f5-427a-bff2-64f6a9b80660"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:00:03 crc kubenswrapper[4815]: I1205 10:00:03.590934 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e42a353-a6f5-427a-bff2-64f6a9b80660-kube-api-access-5s5hd" (OuterVolumeSpecName: "kube-api-access-5s5hd") pod "1e42a353-a6f5-427a-bff2-64f6a9b80660" (UID: "1e42a353-a6f5-427a-bff2-64f6a9b80660"). InnerVolumeSpecName "kube-api-access-5s5hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:00:03 crc kubenswrapper[4815]: I1205 10:00:03.686082 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1e42a353-a6f5-427a-bff2-64f6a9b80660-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:00:03 crc kubenswrapper[4815]: I1205 10:00:03.686574 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5s5hd\" (UniqueName: \"kubernetes.io/projected/1e42a353-a6f5-427a-bff2-64f6a9b80660-kube-api-access-5s5hd\") on node \"crc\" DevicePath \"\"" Dec 05 10:00:03 crc kubenswrapper[4815]: I1205 10:00:03.686889 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1e42a353-a6f5-427a-bff2-64f6a9b80660-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:00:04 crc kubenswrapper[4815]: I1205 10:00:04.096772 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" event={"ID":"1e42a353-a6f5-427a-bff2-64f6a9b80660","Type":"ContainerDied","Data":"71f6d0504c167ddf5850aa1e1ebf1721e5b9a5c5f692f5cac9593d9bbdc6dbad"} Dec 05 10:00:04 crc kubenswrapper[4815]: I1205 10:00:04.096815 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71f6d0504c167ddf5850aa1e1ebf1721e5b9a5c5f692f5cac9593d9bbdc6dbad" Dec 05 10:00:04 crc kubenswrapper[4815]: I1205 10:00:04.096822 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb" Dec 05 10:00:04 crc kubenswrapper[4815]: I1205 10:00:04.569117 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd"] Dec 05 10:00:04 crc kubenswrapper[4815]: I1205 10:00:04.580647 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415435-p2dgd"] Dec 05 10:00:05 crc kubenswrapper[4815]: I1205 10:00:05.431162 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abe6eb7f-9d83-4554-bdb7-1d6e20693873" path="/var/lib/kubelet/pods/abe6eb7f-9d83-4554-bdb7-1d6e20693873/volumes" Dec 05 10:00:09 crc kubenswrapper[4815]: I1205 10:00:09.419293 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 10:00:09 crc kubenswrapper[4815]: E1205 10:00:09.420087 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:00:24 crc kubenswrapper[4815]: I1205 10:00:24.419000 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 10:00:24 crc kubenswrapper[4815]: E1205 10:00:24.419807 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:00:36 crc kubenswrapper[4815]: I1205 10:00:36.419328 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 10:00:36 crc kubenswrapper[4815]: E1205 10:00:36.420291 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:00:51 crc kubenswrapper[4815]: I1205 10:00:51.424865 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 10:00:51 crc kubenswrapper[4815]: E1205 10:00:51.425602 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.188103 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29415481-7x47p"] Dec 05 10:01:00 crc kubenswrapper[4815]: E1205 10:01:00.189309 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e42a353-a6f5-427a-bff2-64f6a9b80660" containerName="collect-profiles" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.189345 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e42a353-a6f5-427a-bff2-64f6a9b80660" containerName="collect-profiles" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.189774 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e42a353-a6f5-427a-bff2-64f6a9b80660" containerName="collect-profiles" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.190726 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.203749 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415481-7x47p"] Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.306325 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-combined-ca-bundle\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.306681 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-config-data\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.306880 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-fernet-keys\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.307001 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t2dq\" (UniqueName: \"kubernetes.io/projected/fba6bed7-9ef2-408d-ac41-04bff23ec468-kube-api-access-4t2dq\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.410100 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-fernet-keys\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.410235 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t2dq\" (UniqueName: \"kubernetes.io/projected/fba6bed7-9ef2-408d-ac41-04bff23ec468-kube-api-access-4t2dq\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.410829 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-combined-ca-bundle\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.410915 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-config-data\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.417338 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-combined-ca-bundle\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.417534 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-fernet-keys\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.421558 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-config-data\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.434104 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t2dq\" (UniqueName: \"kubernetes.io/projected/fba6bed7-9ef2-408d-ac41-04bff23ec468-kube-api-access-4t2dq\") pod \"keystone-cron-29415481-7x47p\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:00 crc kubenswrapper[4815]: I1205 10:01:00.530894 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:01 crc kubenswrapper[4815]: I1205 10:01:01.024915 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29415481-7x47p"] Dec 05 10:01:01 crc kubenswrapper[4815]: I1205 10:01:01.859552 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415481-7x47p" event={"ID":"fba6bed7-9ef2-408d-ac41-04bff23ec468","Type":"ContainerStarted","Data":"664f759339d26d5be2a00376ea20924c519a033116fd6d236ea72d03ec9e1776"} Dec 05 10:01:01 crc kubenswrapper[4815]: I1205 10:01:01.859876 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415481-7x47p" event={"ID":"fba6bed7-9ef2-408d-ac41-04bff23ec468","Type":"ContainerStarted","Data":"fa65f501215b5764dc65802e024a42c1429ca4a8bd16d9c4542c92ddfa8e50b0"} Dec 05 10:01:01 crc kubenswrapper[4815]: I1205 10:01:01.884264 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29415481-7x47p" podStartSLOduration=1.8842373540000001 podStartE2EDuration="1.884237354s" podCreationTimestamp="2025-12-05 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:01:01.881969373 +0000 UTC m=+3260.760576210" watchObservedRunningTime="2025-12-05 10:01:01.884237354 +0000 UTC m=+3260.762844191" Dec 05 10:01:02 crc kubenswrapper[4815]: I1205 10:01:02.519504 4815 scope.go:117] "RemoveContainer" containerID="64b87e99d479d12b7384688d39169bbe50246e09c74baa61049e6bf8be19ad2c" Dec 05 10:01:04 crc kubenswrapper[4815]: I1205 10:01:04.890645 4815 generic.go:334] "Generic (PLEG): container finished" podID="fba6bed7-9ef2-408d-ac41-04bff23ec468" containerID="664f759339d26d5be2a00376ea20924c519a033116fd6d236ea72d03ec9e1776" exitCode=0 Dec 05 10:01:04 crc kubenswrapper[4815]: I1205 10:01:04.890755 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415481-7x47p" event={"ID":"fba6bed7-9ef2-408d-ac41-04bff23ec468","Type":"ContainerDied","Data":"664f759339d26d5be2a00376ea20924c519a033116fd6d236ea72d03ec9e1776"} Dec 05 10:01:05 crc kubenswrapper[4815]: I1205 10:01:05.418721 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 10:01:05 crc kubenswrapper[4815]: E1205 10:01:05.418957 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.296125 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.426012 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-fernet-keys\") pod \"fba6bed7-9ef2-408d-ac41-04bff23ec468\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.426066 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-config-data\") pod \"fba6bed7-9ef2-408d-ac41-04bff23ec468\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.426109 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-combined-ca-bundle\") pod \"fba6bed7-9ef2-408d-ac41-04bff23ec468\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.426136 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t2dq\" (UniqueName: \"kubernetes.io/projected/fba6bed7-9ef2-408d-ac41-04bff23ec468-kube-api-access-4t2dq\") pod \"fba6bed7-9ef2-408d-ac41-04bff23ec468\" (UID: \"fba6bed7-9ef2-408d-ac41-04bff23ec468\") " Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.431751 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "fba6bed7-9ef2-408d-ac41-04bff23ec468" (UID: "fba6bed7-9ef2-408d-ac41-04bff23ec468"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.432036 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fba6bed7-9ef2-408d-ac41-04bff23ec468-kube-api-access-4t2dq" (OuterVolumeSpecName: "kube-api-access-4t2dq") pod "fba6bed7-9ef2-408d-ac41-04bff23ec468" (UID: "fba6bed7-9ef2-408d-ac41-04bff23ec468"). InnerVolumeSpecName "kube-api-access-4t2dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.465773 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fba6bed7-9ef2-408d-ac41-04bff23ec468" (UID: "fba6bed7-9ef2-408d-ac41-04bff23ec468"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.497082 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-config-data" (OuterVolumeSpecName: "config-data") pod "fba6bed7-9ef2-408d-ac41-04bff23ec468" (UID: "fba6bed7-9ef2-408d-ac41-04bff23ec468"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.528699 4815 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.528740 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.528759 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fba6bed7-9ef2-408d-ac41-04bff23ec468-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.528777 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t2dq\" (UniqueName: \"kubernetes.io/projected/fba6bed7-9ef2-408d-ac41-04bff23ec468-kube-api-access-4t2dq\") on node \"crc\" DevicePath \"\"" Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.931665 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29415481-7x47p" event={"ID":"fba6bed7-9ef2-408d-ac41-04bff23ec468","Type":"ContainerDied","Data":"fa65f501215b5764dc65802e024a42c1429ca4a8bd16d9c4542c92ddfa8e50b0"} Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.931746 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa65f501215b5764dc65802e024a42c1429ca4a8bd16d9c4542c92ddfa8e50b0" Dec 05 10:01:06 crc kubenswrapper[4815]: I1205 10:01:06.931868 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29415481-7x47p" Dec 05 10:01:20 crc kubenswrapper[4815]: I1205 10:01:20.418785 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 10:01:20 crc kubenswrapper[4815]: E1205 10:01:20.420012 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:01:33 crc kubenswrapper[4815]: I1205 10:01:33.419272 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 10:01:33 crc kubenswrapper[4815]: E1205 10:01:33.420939 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:01:44 crc kubenswrapper[4815]: I1205 10:01:44.419011 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 10:01:44 crc kubenswrapper[4815]: E1205 10:01:44.419810 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:01:55 crc kubenswrapper[4815]: I1205 10:01:55.419474 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 10:01:56 crc kubenswrapper[4815]: I1205 10:01:56.374206 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"8af74f205ea1ac61f50c412f382831877290dbde40fda4bf5e34911bd0856599"} Dec 05 10:02:02 crc kubenswrapper[4815]: I1205 10:02:02.579421 4815 scope.go:117] "RemoveContainer" containerID="e04cd08e9eb6bc1dfd4063a5defcdeac81e3adbad58c07079ee0b9a6986426ce" Dec 05 10:02:02 crc kubenswrapper[4815]: I1205 10:02:02.618986 4815 scope.go:117] "RemoveContainer" containerID="12532b149458ec3f99551fa4fdb48915f7d527484b0ebebc22f762ca99009352" Dec 05 10:02:02 crc kubenswrapper[4815]: I1205 10:02:02.648788 4815 scope.go:117] "RemoveContainer" containerID="0ad30ebb4a059d7315bd85832102f25a9d54122e34d0601dfb1be9899dbb434f" Dec 05 10:04:20 crc kubenswrapper[4815]: I1205 10:04:20.192792 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:04:20 crc kubenswrapper[4815]: I1205 10:04:20.193410 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:04:22 crc kubenswrapper[4815]: I1205 10:04:22.741286 4815 generic.go:334] "Generic (PLEG): container finished" podID="72f92018-4b16-474a-b5de-8fd124dd857b" containerID="aaa542e1ea30ca7bf409a862fb65a0a5c46f6f53c202c0a75445e1d7c46d1498" exitCode=0 Dec 05 10:04:22 crc kubenswrapper[4815]: I1205 10:04:22.741367 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" event={"ID":"72f92018-4b16-474a-b5de-8fd124dd857b","Type":"ContainerDied","Data":"aaa542e1ea30ca7bf409a862fb65a0a5c46f6f53c202c0a75445e1d7c46d1498"} Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.205236 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.379874 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xc9l6\" (UniqueName: \"kubernetes.io/projected/72f92018-4b16-474a-b5de-8fd124dd857b-kube-api-access-xc9l6\") pod \"72f92018-4b16-474a-b5de-8fd124dd857b\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.380235 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ssh-key\") pod \"72f92018-4b16-474a-b5de-8fd124dd857b\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.380413 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-inventory\") pod \"72f92018-4b16-474a-b5de-8fd124dd857b\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.380596 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-secret-0\") pod \"72f92018-4b16-474a-b5de-8fd124dd857b\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.380648 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-combined-ca-bundle\") pod \"72f92018-4b16-474a-b5de-8fd124dd857b\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.380719 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ceph\") pod \"72f92018-4b16-474a-b5de-8fd124dd857b\" (UID: \"72f92018-4b16-474a-b5de-8fd124dd857b\") " Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.387887 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72f92018-4b16-474a-b5de-8fd124dd857b-kube-api-access-xc9l6" (OuterVolumeSpecName: "kube-api-access-xc9l6") pod "72f92018-4b16-474a-b5de-8fd124dd857b" (UID: "72f92018-4b16-474a-b5de-8fd124dd857b"). InnerVolumeSpecName "kube-api-access-xc9l6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.401684 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "72f92018-4b16-474a-b5de-8fd124dd857b" (UID: "72f92018-4b16-474a-b5de-8fd124dd857b"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.413872 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ceph" (OuterVolumeSpecName: "ceph") pod "72f92018-4b16-474a-b5de-8fd124dd857b" (UID: "72f92018-4b16-474a-b5de-8fd124dd857b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.421542 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-inventory" (OuterVolumeSpecName: "inventory") pod "72f92018-4b16-474a-b5de-8fd124dd857b" (UID: "72f92018-4b16-474a-b5de-8fd124dd857b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.426888 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "72f92018-4b16-474a-b5de-8fd124dd857b" (UID: "72f92018-4b16-474a-b5de-8fd124dd857b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.436035 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "72f92018-4b16-474a-b5de-8fd124dd857b" (UID: "72f92018-4b16-474a-b5de-8fd124dd857b"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.482762 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.482798 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xc9l6\" (UniqueName: \"kubernetes.io/projected/72f92018-4b16-474a-b5de-8fd124dd857b-kube-api-access-xc9l6\") on node \"crc\" DevicePath \"\"" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.482809 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.482817 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.482830 4815 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.482839 4815 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/72f92018-4b16-474a-b5de-8fd124dd857b-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.765245 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" event={"ID":"72f92018-4b16-474a-b5de-8fd124dd857b","Type":"ContainerDied","Data":"feb4a11b92871d9ec65088847a7a5738b2cb1046682cff42b5fbfa105ce57f7c"} Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.765595 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="feb4a11b92871d9ec65088847a7a5738b2cb1046682cff42b5fbfa105ce57f7c" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.765822 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-kknjs" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.883824 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl"] Dec 05 10:04:24 crc kubenswrapper[4815]: E1205 10:04:24.884348 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fba6bed7-9ef2-408d-ac41-04bff23ec468" containerName="keystone-cron" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.884387 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="fba6bed7-9ef2-408d-ac41-04bff23ec468" containerName="keystone-cron" Dec 05 10:04:24 crc kubenswrapper[4815]: E1205 10:04:24.884412 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72f92018-4b16-474a-b5de-8fd124dd857b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.884422 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="72f92018-4b16-474a-b5de-8fd124dd857b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.884778 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="72f92018-4b16-474a-b5de-8fd124dd857b" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.884811 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="fba6bed7-9ef2-408d-ac41-04bff23ec468" containerName="keystone-cron" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.885839 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.889313 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.889760 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.889835 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.889967 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.890196 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.890424 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-jzcrc" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.890597 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.890904 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.892432 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.908778 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl"] Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.991798 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.991862 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.992551 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2k5n\" (UniqueName: \"kubernetes.io/projected/e6418e66-6e27-4292-88c6-fca958ba3665-kube-api-access-d2k5n\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.992676 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.992768 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.992794 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.992827 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.992846 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.992897 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.992952 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:24 crc kubenswrapper[4815]: I1205 10:04:24.992983 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.094840 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2k5n\" (UniqueName: \"kubernetes.io/projected/e6418e66-6e27-4292-88c6-fca958ba3665-kube-api-access-d2k5n\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.094994 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.095030 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.095065 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.095105 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.095133 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.095173 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.095215 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.095257 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.095328 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.095360 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.096318 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.097016 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.099769 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.100286 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.100987 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.101668 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.101806 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.102776 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.102794 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.108998 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.117460 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2k5n\" (UniqueName: \"kubernetes.io/projected/e6418e66-6e27-4292-88c6-fca958ba3665-kube-api-access-d2k5n\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.204713 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.825042 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl"] Dec 05 10:04:25 crc kubenswrapper[4815]: I1205 10:04:25.835537 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:04:26 crc kubenswrapper[4815]: I1205 10:04:26.799264 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" event={"ID":"e6418e66-6e27-4292-88c6-fca958ba3665","Type":"ContainerStarted","Data":"bbbed17c95d9aef3cfaa95ef744a4013fab3bac5e7bdfdf2be1676db2cac4a82"} Dec 05 10:04:26 crc kubenswrapper[4815]: I1205 10:04:26.799674 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" event={"ID":"e6418e66-6e27-4292-88c6-fca958ba3665","Type":"ContainerStarted","Data":"3fb785a35dfd76520e84c917511a2ce376c9f64f1f8d521f1f75f181aad9c6d5"} Dec 05 10:04:26 crc kubenswrapper[4815]: I1205 10:04:26.822102 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" podStartSLOduration=2.263920718 podStartE2EDuration="2.822073473s" podCreationTimestamp="2025-12-05 10:04:24 +0000 UTC" firstStartedPulling="2025-12-05 10:04:25.835260473 +0000 UTC m=+3464.713867310" lastFinishedPulling="2025-12-05 10:04:26.393413238 +0000 UTC m=+3465.272020065" observedRunningTime="2025-12-05 10:04:26.815911875 +0000 UTC m=+3465.694518742" watchObservedRunningTime="2025-12-05 10:04:26.822073473 +0000 UTC m=+3465.700680300" Dec 05 10:04:50 crc kubenswrapper[4815]: I1205 10:04:50.192360 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:04:50 crc kubenswrapper[4815]: I1205 10:04:50.192925 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:05:20 crc kubenswrapper[4815]: I1205 10:05:20.191952 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:05:20 crc kubenswrapper[4815]: I1205 10:05:20.192595 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:05:20 crc kubenswrapper[4815]: I1205 10:05:20.192662 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 10:05:20 crc kubenswrapper[4815]: I1205 10:05:20.193517 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8af74f205ea1ac61f50c412f382831877290dbde40fda4bf5e34911bd0856599"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:05:20 crc kubenswrapper[4815]: I1205 10:05:20.193582 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://8af74f205ea1ac61f50c412f382831877290dbde40fda4bf5e34911bd0856599" gracePeriod=600 Dec 05 10:05:21 crc kubenswrapper[4815]: I1205 10:05:21.323043 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="8af74f205ea1ac61f50c412f382831877290dbde40fda4bf5e34911bd0856599" exitCode=0 Dec 05 10:05:21 crc kubenswrapper[4815]: I1205 10:05:21.323118 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"8af74f205ea1ac61f50c412f382831877290dbde40fda4bf5e34911bd0856599"} Dec 05 10:05:21 crc kubenswrapper[4815]: I1205 10:05:21.323633 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1"} Dec 05 10:05:21 crc kubenswrapper[4815]: I1205 10:05:21.323690 4815 scope.go:117] "RemoveContainer" containerID="6fc96794c8a218a55c01b6ba2c604be35ddee23c667ebbeb194a5e00723bd754" Dec 05 10:06:02 crc kubenswrapper[4815]: I1205 10:06:02.943315 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5rcbb"] Dec 05 10:06:02 crc kubenswrapper[4815]: I1205 10:06:02.947631 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:02 crc kubenswrapper[4815]: I1205 10:06:02.966553 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rcbb"] Dec 05 10:06:03 crc kubenswrapper[4815]: I1205 10:06:03.126357 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-catalog-content\") pod \"community-operators-5rcbb\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:03 crc kubenswrapper[4815]: I1205 10:06:03.126419 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-utilities\") pod \"community-operators-5rcbb\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:03 crc kubenswrapper[4815]: I1205 10:06:03.126759 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6zph\" (UniqueName: \"kubernetes.io/projected/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-kube-api-access-r6zph\") pod \"community-operators-5rcbb\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:03 crc kubenswrapper[4815]: I1205 10:06:03.228378 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6zph\" (UniqueName: \"kubernetes.io/projected/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-kube-api-access-r6zph\") pod \"community-operators-5rcbb\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:03 crc kubenswrapper[4815]: I1205 10:06:03.228544 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-catalog-content\") pod \"community-operators-5rcbb\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:03 crc kubenswrapper[4815]: I1205 10:06:03.228579 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-utilities\") pod \"community-operators-5rcbb\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:03 crc kubenswrapper[4815]: I1205 10:06:03.229087 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-utilities\") pod \"community-operators-5rcbb\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:03 crc kubenswrapper[4815]: I1205 10:06:03.229181 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-catalog-content\") pod \"community-operators-5rcbb\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:03 crc kubenswrapper[4815]: I1205 10:06:03.258373 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6zph\" (UniqueName: \"kubernetes.io/projected/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-kube-api-access-r6zph\") pod \"community-operators-5rcbb\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:03 crc kubenswrapper[4815]: I1205 10:06:03.286561 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:03 crc kubenswrapper[4815]: I1205 10:06:03.907994 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5rcbb"] Dec 05 10:06:03 crc kubenswrapper[4815]: W1205 10:06:03.918332 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda1a75df9_53c7_4a9f_b546_5779d05b1a9e.slice/crio-9db3c833de706619cc4959febecf2a2d6a0f622aabc1b431d3ce6292814933d0 WatchSource:0}: Error finding container 9db3c833de706619cc4959febecf2a2d6a0f622aabc1b431d3ce6292814933d0: Status 404 returned error can't find the container with id 9db3c833de706619cc4959febecf2a2d6a0f622aabc1b431d3ce6292814933d0 Dec 05 10:06:04 crc kubenswrapper[4815]: I1205 10:06:04.916778 4815 generic.go:334] "Generic (PLEG): container finished" podID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" containerID="22723edb8d96c8aaa5eeccf89922dea04844733454f15738d5f8e4873cc094e6" exitCode=0 Dec 05 10:06:04 crc kubenswrapper[4815]: I1205 10:06:04.917019 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rcbb" event={"ID":"a1a75df9-53c7-4a9f-b546-5779d05b1a9e","Type":"ContainerDied","Data":"22723edb8d96c8aaa5eeccf89922dea04844733454f15738d5f8e4873cc094e6"} Dec 05 10:06:04 crc kubenswrapper[4815]: I1205 10:06:04.917050 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rcbb" event={"ID":"a1a75df9-53c7-4a9f-b546-5779d05b1a9e","Type":"ContainerStarted","Data":"9db3c833de706619cc4959febecf2a2d6a0f622aabc1b431d3ce6292814933d0"} Dec 05 10:06:07 crc kubenswrapper[4815]: I1205 10:06:07.950187 4815 generic.go:334] "Generic (PLEG): container finished" podID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" containerID="d2ff2cbe79f10f6a45fffdbad9e75ebb3361324b01a91276004e1826f9ee844b" exitCode=0 Dec 05 10:06:07 crc kubenswrapper[4815]: I1205 10:06:07.950315 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rcbb" event={"ID":"a1a75df9-53c7-4a9f-b546-5779d05b1a9e","Type":"ContainerDied","Data":"d2ff2cbe79f10f6a45fffdbad9e75ebb3361324b01a91276004e1826f9ee844b"} Dec 05 10:06:09 crc kubenswrapper[4815]: I1205 10:06:09.971310 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rcbb" event={"ID":"a1a75df9-53c7-4a9f-b546-5779d05b1a9e","Type":"ContainerStarted","Data":"66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1"} Dec 05 10:06:09 crc kubenswrapper[4815]: I1205 10:06:09.994884 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5rcbb" podStartSLOduration=3.888022192 podStartE2EDuration="7.994842036s" podCreationTimestamp="2025-12-05 10:06:02 +0000 UTC" firstStartedPulling="2025-12-05 10:06:04.919481296 +0000 UTC m=+3563.798088133" lastFinishedPulling="2025-12-05 10:06:09.02630114 +0000 UTC m=+3567.904907977" observedRunningTime="2025-12-05 10:06:09.991798053 +0000 UTC m=+3568.870404890" watchObservedRunningTime="2025-12-05 10:06:09.994842036 +0000 UTC m=+3568.873448883" Dec 05 10:06:13 crc kubenswrapper[4815]: I1205 10:06:13.287576 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:13 crc kubenswrapper[4815]: I1205 10:06:13.289201 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:13 crc kubenswrapper[4815]: I1205 10:06:13.335753 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:14 crc kubenswrapper[4815]: I1205 10:06:14.130591 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:15 crc kubenswrapper[4815]: I1205 10:06:15.078291 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5rcbb"] Dec 05 10:06:17 crc kubenswrapper[4815]: I1205 10:06:17.037483 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5rcbb" podUID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" containerName="registry-server" containerID="cri-o://66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1" gracePeriod=2 Dec 05 10:06:17 crc kubenswrapper[4815]: I1205 10:06:17.546819 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:17 crc kubenswrapper[4815]: I1205 10:06:17.599009 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6zph\" (UniqueName: \"kubernetes.io/projected/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-kube-api-access-r6zph\") pod \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " Dec 05 10:06:17 crc kubenswrapper[4815]: I1205 10:06:17.599058 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-catalog-content\") pod \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " Dec 05 10:06:17 crc kubenswrapper[4815]: I1205 10:06:17.599097 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-utilities\") pod \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\" (UID: \"a1a75df9-53c7-4a9f-b546-5779d05b1a9e\") " Dec 05 10:06:17 crc kubenswrapper[4815]: I1205 10:06:17.600147 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-utilities" (OuterVolumeSpecName: "utilities") pod "a1a75df9-53c7-4a9f-b546-5779d05b1a9e" (UID: "a1a75df9-53c7-4a9f-b546-5779d05b1a9e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:06:17 crc kubenswrapper[4815]: I1205 10:06:17.619108 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-kube-api-access-r6zph" (OuterVolumeSpecName: "kube-api-access-r6zph") pod "a1a75df9-53c7-4a9f-b546-5779d05b1a9e" (UID: "a1a75df9-53c7-4a9f-b546-5779d05b1a9e"). InnerVolumeSpecName "kube-api-access-r6zph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:06:17 crc kubenswrapper[4815]: I1205 10:06:17.655157 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1a75df9-53c7-4a9f-b546-5779d05b1a9e" (UID: "a1a75df9-53c7-4a9f-b546-5779d05b1a9e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:06:17 crc kubenswrapper[4815]: I1205 10:06:17.700313 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:06:17 crc kubenswrapper[4815]: I1205 10:06:17.700346 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6zph\" (UniqueName: \"kubernetes.io/projected/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-kube-api-access-r6zph\") on node \"crc\" DevicePath \"\"" Dec 05 10:06:17 crc kubenswrapper[4815]: I1205 10:06:17.700356 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1a75df9-53c7-4a9f-b546-5779d05b1a9e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.047071 4815 generic.go:334] "Generic (PLEG): container finished" podID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" containerID="66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1" exitCode=0 Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.047112 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rcbb" event={"ID":"a1a75df9-53c7-4a9f-b546-5779d05b1a9e","Type":"ContainerDied","Data":"66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1"} Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.047120 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5rcbb" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.047143 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5rcbb" event={"ID":"a1a75df9-53c7-4a9f-b546-5779d05b1a9e","Type":"ContainerDied","Data":"9db3c833de706619cc4959febecf2a2d6a0f622aabc1b431d3ce6292814933d0"} Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.047185 4815 scope.go:117] "RemoveContainer" containerID="66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.067599 4815 scope.go:117] "RemoveContainer" containerID="d2ff2cbe79f10f6a45fffdbad9e75ebb3361324b01a91276004e1826f9ee844b" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.087482 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5rcbb"] Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.101839 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5rcbb"] Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.104109 4815 scope.go:117] "RemoveContainer" containerID="22723edb8d96c8aaa5eeccf89922dea04844733454f15738d5f8e4873cc094e6" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.147654 4815 scope.go:117] "RemoveContainer" containerID="66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1" Dec 05 10:06:18 crc kubenswrapper[4815]: E1205 10:06:18.148057 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1\": container with ID starting with 66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1 not found: ID does not exist" containerID="66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.148120 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1"} err="failed to get container status \"66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1\": rpc error: code = NotFound desc = could not find container \"66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1\": container with ID starting with 66ef58b8f3c7b5b12a83ac490323c94a8ee95a034afc13cebc3041d5bff846d1 not found: ID does not exist" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.148147 4815 scope.go:117] "RemoveContainer" containerID="d2ff2cbe79f10f6a45fffdbad9e75ebb3361324b01a91276004e1826f9ee844b" Dec 05 10:06:18 crc kubenswrapper[4815]: E1205 10:06:18.148346 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2ff2cbe79f10f6a45fffdbad9e75ebb3361324b01a91276004e1826f9ee844b\": container with ID starting with d2ff2cbe79f10f6a45fffdbad9e75ebb3361324b01a91276004e1826f9ee844b not found: ID does not exist" containerID="d2ff2cbe79f10f6a45fffdbad9e75ebb3361324b01a91276004e1826f9ee844b" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.148401 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2ff2cbe79f10f6a45fffdbad9e75ebb3361324b01a91276004e1826f9ee844b"} err="failed to get container status \"d2ff2cbe79f10f6a45fffdbad9e75ebb3361324b01a91276004e1826f9ee844b\": rpc error: code = NotFound desc = could not find container \"d2ff2cbe79f10f6a45fffdbad9e75ebb3361324b01a91276004e1826f9ee844b\": container with ID starting with d2ff2cbe79f10f6a45fffdbad9e75ebb3361324b01a91276004e1826f9ee844b not found: ID does not exist" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.148420 4815 scope.go:117] "RemoveContainer" containerID="22723edb8d96c8aaa5eeccf89922dea04844733454f15738d5f8e4873cc094e6" Dec 05 10:06:18 crc kubenswrapper[4815]: E1205 10:06:18.148862 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22723edb8d96c8aaa5eeccf89922dea04844733454f15738d5f8e4873cc094e6\": container with ID starting with 22723edb8d96c8aaa5eeccf89922dea04844733454f15738d5f8e4873cc094e6 not found: ID does not exist" containerID="22723edb8d96c8aaa5eeccf89922dea04844733454f15738d5f8e4873cc094e6" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.148888 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22723edb8d96c8aaa5eeccf89922dea04844733454f15738d5f8e4873cc094e6"} err="failed to get container status \"22723edb8d96c8aaa5eeccf89922dea04844733454f15738d5f8e4873cc094e6\": rpc error: code = NotFound desc = could not find container \"22723edb8d96c8aaa5eeccf89922dea04844733454f15738d5f8e4873cc094e6\": container with ID starting with 22723edb8d96c8aaa5eeccf89922dea04844733454f15738d5f8e4873cc094e6 not found: ID does not exist" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.690393 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wqndm"] Dec 05 10:06:18 crc kubenswrapper[4815]: E1205 10:06:18.691098 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" containerName="extract-utilities" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.691135 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" containerName="extract-utilities" Dec 05 10:06:18 crc kubenswrapper[4815]: E1205 10:06:18.691148 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" containerName="registry-server" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.691155 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" containerName="registry-server" Dec 05 10:06:18 crc kubenswrapper[4815]: E1205 10:06:18.691170 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" containerName="extract-content" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.691176 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" containerName="extract-content" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.691372 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" containerName="registry-server" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.692824 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.708374 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqndm"] Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.827458 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-444tm\" (UniqueName: \"kubernetes.io/projected/88250b2d-4867-4f08-97dd-9242504f0bdf-kube-api-access-444tm\") pod \"redhat-marketplace-wqndm\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.828141 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-utilities\") pod \"redhat-marketplace-wqndm\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.828212 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-catalog-content\") pod \"redhat-marketplace-wqndm\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.930795 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-utilities\") pod \"redhat-marketplace-wqndm\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.930874 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-catalog-content\") pod \"redhat-marketplace-wqndm\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.931057 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-444tm\" (UniqueName: \"kubernetes.io/projected/88250b2d-4867-4f08-97dd-9242504f0bdf-kube-api-access-444tm\") pod \"redhat-marketplace-wqndm\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.931317 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-utilities\") pod \"redhat-marketplace-wqndm\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.931317 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-catalog-content\") pod \"redhat-marketplace-wqndm\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:18 crc kubenswrapper[4815]: I1205 10:06:18.949333 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-444tm\" (UniqueName: \"kubernetes.io/projected/88250b2d-4867-4f08-97dd-9242504f0bdf-kube-api-access-444tm\") pod \"redhat-marketplace-wqndm\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:19 crc kubenswrapper[4815]: I1205 10:06:19.010120 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:19 crc kubenswrapper[4815]: I1205 10:06:19.431960 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1a75df9-53c7-4a9f-b546-5779d05b1a9e" path="/var/lib/kubelet/pods/a1a75df9-53c7-4a9f-b546-5779d05b1a9e/volumes" Dec 05 10:06:19 crc kubenswrapper[4815]: I1205 10:06:19.687873 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqndm"] Dec 05 10:06:20 crc kubenswrapper[4815]: I1205 10:06:20.099546 4815 generic.go:334] "Generic (PLEG): container finished" podID="88250b2d-4867-4f08-97dd-9242504f0bdf" containerID="226d35fd6a44e7dba4b8db3f245cd30505cd83a8ea434e2bf98cfaae1b7166ca" exitCode=0 Dec 05 10:06:20 crc kubenswrapper[4815]: I1205 10:06:20.099643 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqndm" event={"ID":"88250b2d-4867-4f08-97dd-9242504f0bdf","Type":"ContainerDied","Data":"226d35fd6a44e7dba4b8db3f245cd30505cd83a8ea434e2bf98cfaae1b7166ca"} Dec 05 10:06:20 crc kubenswrapper[4815]: I1205 10:06:20.099706 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqndm" event={"ID":"88250b2d-4867-4f08-97dd-9242504f0bdf","Type":"ContainerStarted","Data":"60397ab755bcdac4f0aa31337e06b7225d32630fcf581be58b45d2eb9ed3d577"} Dec 05 10:06:22 crc kubenswrapper[4815]: I1205 10:06:22.118741 4815 generic.go:334] "Generic (PLEG): container finished" podID="88250b2d-4867-4f08-97dd-9242504f0bdf" containerID="014387ce1116a9f016f9ad76314dfc5e9f1264f0e8f23a054f348b1c36f08dc9" exitCode=0 Dec 05 10:06:22 crc kubenswrapper[4815]: I1205 10:06:22.118945 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqndm" event={"ID":"88250b2d-4867-4f08-97dd-9242504f0bdf","Type":"ContainerDied","Data":"014387ce1116a9f016f9ad76314dfc5e9f1264f0e8f23a054f348b1c36f08dc9"} Dec 05 10:06:22 crc kubenswrapper[4815]: I1205 10:06:22.881021 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-967rh"] Dec 05 10:06:22 crc kubenswrapper[4815]: I1205 10:06:22.883079 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:22 crc kubenswrapper[4815]: I1205 10:06:22.893171 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-967rh"] Dec 05 10:06:23 crc kubenswrapper[4815]: I1205 10:06:23.043673 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gln58\" (UniqueName: \"kubernetes.io/projected/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-kube-api-access-gln58\") pod \"certified-operators-967rh\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:23 crc kubenswrapper[4815]: I1205 10:06:23.044231 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-utilities\") pod \"certified-operators-967rh\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:23 crc kubenswrapper[4815]: I1205 10:06:23.044267 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-catalog-content\") pod \"certified-operators-967rh\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:23 crc kubenswrapper[4815]: I1205 10:06:23.145903 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-utilities\") pod \"certified-operators-967rh\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:23 crc kubenswrapper[4815]: I1205 10:06:23.146975 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-catalog-content\") pod \"certified-operators-967rh\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:23 crc kubenswrapper[4815]: I1205 10:06:23.147284 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-catalog-content\") pod \"certified-operators-967rh\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:23 crc kubenswrapper[4815]: I1205 10:06:23.147282 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gln58\" (UniqueName: \"kubernetes.io/projected/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-kube-api-access-gln58\") pod \"certified-operators-967rh\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:23 crc kubenswrapper[4815]: I1205 10:06:23.146388 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-utilities\") pod \"certified-operators-967rh\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:23 crc kubenswrapper[4815]: I1205 10:06:23.169303 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gln58\" (UniqueName: \"kubernetes.io/projected/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-kube-api-access-gln58\") pod \"certified-operators-967rh\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:23 crc kubenswrapper[4815]: I1205 10:06:23.253921 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:24 crc kubenswrapper[4815]: I1205 10:06:23.867429 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-967rh"] Dec 05 10:06:24 crc kubenswrapper[4815]: I1205 10:06:24.156703 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-967rh" event={"ID":"198ea20b-fdfd-4928-9dfd-e1b77af7ece0","Type":"ContainerStarted","Data":"ebe0558b67ef09e9ddf53312a3de04a3189cc5da066a7aa3ffb27ed124d46ae7"} Dec 05 10:06:25 crc kubenswrapper[4815]: I1205 10:06:25.171109 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqndm" event={"ID":"88250b2d-4867-4f08-97dd-9242504f0bdf","Type":"ContainerStarted","Data":"a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208"} Dec 05 10:06:25 crc kubenswrapper[4815]: I1205 10:06:25.174588 4815 generic.go:334] "Generic (PLEG): container finished" podID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" containerID="a4d4278f093ea135acbdf5a951527860b8a80ee888cfb11ff84da99d02a03ac7" exitCode=0 Dec 05 10:06:25 crc kubenswrapper[4815]: I1205 10:06:25.174657 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-967rh" event={"ID":"198ea20b-fdfd-4928-9dfd-e1b77af7ece0","Type":"ContainerDied","Data":"a4d4278f093ea135acbdf5a951527860b8a80ee888cfb11ff84da99d02a03ac7"} Dec 05 10:06:25 crc kubenswrapper[4815]: I1205 10:06:25.197883 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wqndm" podStartSLOduration=3.71849881 podStartE2EDuration="7.1978546s" podCreationTimestamp="2025-12-05 10:06:18 +0000 UTC" firstStartedPulling="2025-12-05 10:06:20.101774923 +0000 UTC m=+3578.980381760" lastFinishedPulling="2025-12-05 10:06:23.581130713 +0000 UTC m=+3582.459737550" observedRunningTime="2025-12-05 10:06:25.191367843 +0000 UTC m=+3584.069974700" watchObservedRunningTime="2025-12-05 10:06:25.1978546 +0000 UTC m=+3584.076461437" Dec 05 10:06:27 crc kubenswrapper[4815]: I1205 10:06:27.193406 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-967rh" event={"ID":"198ea20b-fdfd-4928-9dfd-e1b77af7ece0","Type":"ContainerStarted","Data":"dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf"} Dec 05 10:06:29 crc kubenswrapper[4815]: I1205 10:06:29.010416 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:29 crc kubenswrapper[4815]: I1205 10:06:29.010988 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:29 crc kubenswrapper[4815]: I1205 10:06:29.059430 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:29 crc kubenswrapper[4815]: I1205 10:06:29.216873 4815 generic.go:334] "Generic (PLEG): container finished" podID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" containerID="dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf" exitCode=0 Dec 05 10:06:29 crc kubenswrapper[4815]: I1205 10:06:29.216937 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-967rh" event={"ID":"198ea20b-fdfd-4928-9dfd-e1b77af7ece0","Type":"ContainerDied","Data":"dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf"} Dec 05 10:06:29 crc kubenswrapper[4815]: I1205 10:06:29.271345 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:30 crc kubenswrapper[4815]: I1205 10:06:30.227707 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-967rh" event={"ID":"198ea20b-fdfd-4928-9dfd-e1b77af7ece0","Type":"ContainerStarted","Data":"76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9"} Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.258683 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-967rh" podStartSLOduration=4.606230995 podStartE2EDuration="9.25866367s" podCreationTimestamp="2025-12-05 10:06:22 +0000 UTC" firstStartedPulling="2025-12-05 10:06:25.176023645 +0000 UTC m=+3584.054630482" lastFinishedPulling="2025-12-05 10:06:29.82845632 +0000 UTC m=+3588.707063157" observedRunningTime="2025-12-05 10:06:31.254960889 +0000 UTC m=+3590.133567756" watchObservedRunningTime="2025-12-05 10:06:31.25866367 +0000 UTC m=+3590.137270517" Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.278736 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqndm"] Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.279340 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wqndm" podUID="88250b2d-4867-4f08-97dd-9242504f0bdf" containerName="registry-server" containerID="cri-o://a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208" gracePeriod=2 Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.713370 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.761779 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-444tm\" (UniqueName: \"kubernetes.io/projected/88250b2d-4867-4f08-97dd-9242504f0bdf-kube-api-access-444tm\") pod \"88250b2d-4867-4f08-97dd-9242504f0bdf\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.761950 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-catalog-content\") pod \"88250b2d-4867-4f08-97dd-9242504f0bdf\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.762030 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-utilities\") pod \"88250b2d-4867-4f08-97dd-9242504f0bdf\" (UID: \"88250b2d-4867-4f08-97dd-9242504f0bdf\") " Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.762957 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-utilities" (OuterVolumeSpecName: "utilities") pod "88250b2d-4867-4f08-97dd-9242504f0bdf" (UID: "88250b2d-4867-4f08-97dd-9242504f0bdf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.779533 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88250b2d-4867-4f08-97dd-9242504f0bdf" (UID: "88250b2d-4867-4f08-97dd-9242504f0bdf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.781418 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88250b2d-4867-4f08-97dd-9242504f0bdf-kube-api-access-444tm" (OuterVolumeSpecName: "kube-api-access-444tm") pod "88250b2d-4867-4f08-97dd-9242504f0bdf" (UID: "88250b2d-4867-4f08-97dd-9242504f0bdf"). InnerVolumeSpecName "kube-api-access-444tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.863861 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-444tm\" (UniqueName: \"kubernetes.io/projected/88250b2d-4867-4f08-97dd-9242504f0bdf-kube-api-access-444tm\") on node \"crc\" DevicePath \"\"" Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.864080 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:06:31 crc kubenswrapper[4815]: I1205 10:06:31.864089 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88250b2d-4867-4f08-97dd-9242504f0bdf-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.247156 4815 generic.go:334] "Generic (PLEG): container finished" podID="88250b2d-4867-4f08-97dd-9242504f0bdf" containerID="a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208" exitCode=0 Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.247200 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqndm" event={"ID":"88250b2d-4867-4f08-97dd-9242504f0bdf","Type":"ContainerDied","Data":"a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208"} Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.247226 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wqndm" event={"ID":"88250b2d-4867-4f08-97dd-9242504f0bdf","Type":"ContainerDied","Data":"60397ab755bcdac4f0aa31337e06b7225d32630fcf581be58b45d2eb9ed3d577"} Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.247242 4815 scope.go:117] "RemoveContainer" containerID="a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208" Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.247732 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wqndm" Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.270764 4815 scope.go:117] "RemoveContainer" containerID="014387ce1116a9f016f9ad76314dfc5e9f1264f0e8f23a054f348b1c36f08dc9" Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.305429 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqndm"] Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.306504 4815 scope.go:117] "RemoveContainer" containerID="226d35fd6a44e7dba4b8db3f245cd30505cd83a8ea434e2bf98cfaae1b7166ca" Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.306803 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wqndm"] Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.377931 4815 scope.go:117] "RemoveContainer" containerID="a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208" Dec 05 10:06:32 crc kubenswrapper[4815]: E1205 10:06:32.378660 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208\": container with ID starting with a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208 not found: ID does not exist" containerID="a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208" Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.378795 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208"} err="failed to get container status \"a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208\": rpc error: code = NotFound desc = could not find container \"a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208\": container with ID starting with a2d76f7e0645ca0a6d0a354492e1327b667dbe40494ed5b702498bae13aad208 not found: ID does not exist" Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.378886 4815 scope.go:117] "RemoveContainer" containerID="014387ce1116a9f016f9ad76314dfc5e9f1264f0e8f23a054f348b1c36f08dc9" Dec 05 10:06:32 crc kubenswrapper[4815]: E1205 10:06:32.379336 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"014387ce1116a9f016f9ad76314dfc5e9f1264f0e8f23a054f348b1c36f08dc9\": container with ID starting with 014387ce1116a9f016f9ad76314dfc5e9f1264f0e8f23a054f348b1c36f08dc9 not found: ID does not exist" containerID="014387ce1116a9f016f9ad76314dfc5e9f1264f0e8f23a054f348b1c36f08dc9" Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.379436 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"014387ce1116a9f016f9ad76314dfc5e9f1264f0e8f23a054f348b1c36f08dc9"} err="failed to get container status \"014387ce1116a9f016f9ad76314dfc5e9f1264f0e8f23a054f348b1c36f08dc9\": rpc error: code = NotFound desc = could not find container \"014387ce1116a9f016f9ad76314dfc5e9f1264f0e8f23a054f348b1c36f08dc9\": container with ID starting with 014387ce1116a9f016f9ad76314dfc5e9f1264f0e8f23a054f348b1c36f08dc9 not found: ID does not exist" Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.379537 4815 scope.go:117] "RemoveContainer" containerID="226d35fd6a44e7dba4b8db3f245cd30505cd83a8ea434e2bf98cfaae1b7166ca" Dec 05 10:06:32 crc kubenswrapper[4815]: E1205 10:06:32.379873 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"226d35fd6a44e7dba4b8db3f245cd30505cd83a8ea434e2bf98cfaae1b7166ca\": container with ID starting with 226d35fd6a44e7dba4b8db3f245cd30505cd83a8ea434e2bf98cfaae1b7166ca not found: ID does not exist" containerID="226d35fd6a44e7dba4b8db3f245cd30505cd83a8ea434e2bf98cfaae1b7166ca" Dec 05 10:06:32 crc kubenswrapper[4815]: I1205 10:06:32.379969 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"226d35fd6a44e7dba4b8db3f245cd30505cd83a8ea434e2bf98cfaae1b7166ca"} err="failed to get container status \"226d35fd6a44e7dba4b8db3f245cd30505cd83a8ea434e2bf98cfaae1b7166ca\": rpc error: code = NotFound desc = could not find container \"226d35fd6a44e7dba4b8db3f245cd30505cd83a8ea434e2bf98cfaae1b7166ca\": container with ID starting with 226d35fd6a44e7dba4b8db3f245cd30505cd83a8ea434e2bf98cfaae1b7166ca not found: ID does not exist" Dec 05 10:06:33 crc kubenswrapper[4815]: I1205 10:06:33.254734 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:33 crc kubenswrapper[4815]: I1205 10:06:33.254990 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:33 crc kubenswrapper[4815]: I1205 10:06:33.302707 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:33 crc kubenswrapper[4815]: I1205 10:06:33.428836 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88250b2d-4867-4f08-97dd-9242504f0bdf" path="/var/lib/kubelet/pods/88250b2d-4867-4f08-97dd-9242504f0bdf/volumes" Dec 05 10:06:43 crc kubenswrapper[4815]: I1205 10:06:43.304940 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:43 crc kubenswrapper[4815]: I1205 10:06:43.364859 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-967rh"] Dec 05 10:06:43 crc kubenswrapper[4815]: I1205 10:06:43.365532 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-967rh" podUID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" containerName="registry-server" containerID="cri-o://76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9" gracePeriod=2 Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.322533 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.377587 4815 generic.go:334] "Generic (PLEG): container finished" podID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" containerID="76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9" exitCode=0 Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.377649 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-967rh" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.377697 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-967rh" event={"ID":"198ea20b-fdfd-4928-9dfd-e1b77af7ece0","Type":"ContainerDied","Data":"76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9"} Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.378038 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-967rh" event={"ID":"198ea20b-fdfd-4928-9dfd-e1b77af7ece0","Type":"ContainerDied","Data":"ebe0558b67ef09e9ddf53312a3de04a3189cc5da066a7aa3ffb27ed124d46ae7"} Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.378064 4815 scope.go:117] "RemoveContainer" containerID="76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.417145 4815 scope.go:117] "RemoveContainer" containerID="dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.440040 4815 scope.go:117] "RemoveContainer" containerID="a4d4278f093ea135acbdf5a951527860b8a80ee888cfb11ff84da99d02a03ac7" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.459424 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-catalog-content\") pod \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.459479 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gln58\" (UniqueName: \"kubernetes.io/projected/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-kube-api-access-gln58\") pod \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.459609 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-utilities\") pod \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\" (UID: \"198ea20b-fdfd-4928-9dfd-e1b77af7ece0\") " Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.461960 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-utilities" (OuterVolumeSpecName: "utilities") pod "198ea20b-fdfd-4928-9dfd-e1b77af7ece0" (UID: "198ea20b-fdfd-4928-9dfd-e1b77af7ece0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.473137 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-kube-api-access-gln58" (OuterVolumeSpecName: "kube-api-access-gln58") pod "198ea20b-fdfd-4928-9dfd-e1b77af7ece0" (UID: "198ea20b-fdfd-4928-9dfd-e1b77af7ece0"). InnerVolumeSpecName "kube-api-access-gln58". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.477452 4815 scope.go:117] "RemoveContainer" containerID="76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9" Dec 05 10:06:44 crc kubenswrapper[4815]: E1205 10:06:44.478015 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9\": container with ID starting with 76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9 not found: ID does not exist" containerID="76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.478078 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9"} err="failed to get container status \"76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9\": rpc error: code = NotFound desc = could not find container \"76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9\": container with ID starting with 76b0f03e24a2a4e3c2b5277727f1e093f291dc98a15df79049e6f107648b8fb9 not found: ID does not exist" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.478108 4815 scope.go:117] "RemoveContainer" containerID="dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf" Dec 05 10:06:44 crc kubenswrapper[4815]: E1205 10:06:44.478412 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf\": container with ID starting with dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf not found: ID does not exist" containerID="dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.478449 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf"} err="failed to get container status \"dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf\": rpc error: code = NotFound desc = could not find container \"dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf\": container with ID starting with dc052c275c2651bc46ed8e4c19616c0d230d251045d35c97d1831747f16e0caf not found: ID does not exist" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.478463 4815 scope.go:117] "RemoveContainer" containerID="a4d4278f093ea135acbdf5a951527860b8a80ee888cfb11ff84da99d02a03ac7" Dec 05 10:06:44 crc kubenswrapper[4815]: E1205 10:06:44.479176 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4d4278f093ea135acbdf5a951527860b8a80ee888cfb11ff84da99d02a03ac7\": container with ID starting with a4d4278f093ea135acbdf5a951527860b8a80ee888cfb11ff84da99d02a03ac7 not found: ID does not exist" containerID="a4d4278f093ea135acbdf5a951527860b8a80ee888cfb11ff84da99d02a03ac7" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.479222 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4d4278f093ea135acbdf5a951527860b8a80ee888cfb11ff84da99d02a03ac7"} err="failed to get container status \"a4d4278f093ea135acbdf5a951527860b8a80ee888cfb11ff84da99d02a03ac7\": rpc error: code = NotFound desc = could not find container \"a4d4278f093ea135acbdf5a951527860b8a80ee888cfb11ff84da99d02a03ac7\": container with ID starting with a4d4278f093ea135acbdf5a951527860b8a80ee888cfb11ff84da99d02a03ac7 not found: ID does not exist" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.510342 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "198ea20b-fdfd-4928-9dfd-e1b77af7ece0" (UID: "198ea20b-fdfd-4928-9dfd-e1b77af7ece0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.561938 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.561973 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gln58\" (UniqueName: \"kubernetes.io/projected/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-kube-api-access-gln58\") on node \"crc\" DevicePath \"\"" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.561985 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/198ea20b-fdfd-4928-9dfd-e1b77af7ece0-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.715331 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-967rh"] Dec 05 10:06:44 crc kubenswrapper[4815]: I1205 10:06:44.725746 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-967rh"] Dec 05 10:06:45 crc kubenswrapper[4815]: I1205 10:06:45.428926 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" path="/var/lib/kubelet/pods/198ea20b-fdfd-4928-9dfd-e1b77af7ece0/volumes" Dec 05 10:07:20 crc kubenswrapper[4815]: I1205 10:07:20.192064 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:07:20 crc kubenswrapper[4815]: I1205 10:07:20.192637 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:07:42 crc kubenswrapper[4815]: I1205 10:07:42.971988 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tgjw2"] Dec 05 10:07:42 crc kubenswrapper[4815]: E1205 10:07:42.973164 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" containerName="extract-utilities" Dec 05 10:07:42 crc kubenswrapper[4815]: I1205 10:07:42.973197 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" containerName="extract-utilities" Dec 05 10:07:42 crc kubenswrapper[4815]: E1205 10:07:42.973221 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" containerName="registry-server" Dec 05 10:07:42 crc kubenswrapper[4815]: I1205 10:07:42.973231 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" containerName="registry-server" Dec 05 10:07:42 crc kubenswrapper[4815]: E1205 10:07:42.973247 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88250b2d-4867-4f08-97dd-9242504f0bdf" containerName="extract-utilities" Dec 05 10:07:42 crc kubenswrapper[4815]: I1205 10:07:42.973255 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="88250b2d-4867-4f08-97dd-9242504f0bdf" containerName="extract-utilities" Dec 05 10:07:42 crc kubenswrapper[4815]: E1205 10:07:42.973322 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" containerName="extract-content" Dec 05 10:07:42 crc kubenswrapper[4815]: I1205 10:07:42.973332 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" containerName="extract-content" Dec 05 10:07:42 crc kubenswrapper[4815]: E1205 10:07:42.973360 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88250b2d-4867-4f08-97dd-9242504f0bdf" containerName="extract-content" Dec 05 10:07:42 crc kubenswrapper[4815]: I1205 10:07:42.973368 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="88250b2d-4867-4f08-97dd-9242504f0bdf" containerName="extract-content" Dec 05 10:07:42 crc kubenswrapper[4815]: E1205 10:07:42.973390 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88250b2d-4867-4f08-97dd-9242504f0bdf" containerName="registry-server" Dec 05 10:07:42 crc kubenswrapper[4815]: I1205 10:07:42.973398 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="88250b2d-4867-4f08-97dd-9242504f0bdf" containerName="registry-server" Dec 05 10:07:42 crc kubenswrapper[4815]: I1205 10:07:42.973729 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="198ea20b-fdfd-4928-9dfd-e1b77af7ece0" containerName="registry-server" Dec 05 10:07:42 crc kubenswrapper[4815]: I1205 10:07:42.973754 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="88250b2d-4867-4f08-97dd-9242504f0bdf" containerName="registry-server" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.003082 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tgjw2"] Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.003207 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.122662 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj9th\" (UniqueName: \"kubernetes.io/projected/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-kube-api-access-gj9th\") pod \"redhat-operators-tgjw2\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.122706 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-utilities\") pod \"redhat-operators-tgjw2\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.122752 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-catalog-content\") pod \"redhat-operators-tgjw2\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.224126 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj9th\" (UniqueName: \"kubernetes.io/projected/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-kube-api-access-gj9th\") pod \"redhat-operators-tgjw2\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.224177 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-utilities\") pod \"redhat-operators-tgjw2\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.224364 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-catalog-content\") pod \"redhat-operators-tgjw2\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.224819 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-utilities\") pod \"redhat-operators-tgjw2\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.224860 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-catalog-content\") pod \"redhat-operators-tgjw2\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.246227 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj9th\" (UniqueName: \"kubernetes.io/projected/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-kube-api-access-gj9th\") pod \"redhat-operators-tgjw2\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.328595 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:43 crc kubenswrapper[4815]: I1205 10:07:43.912657 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tgjw2"] Dec 05 10:07:44 crc kubenswrapper[4815]: I1205 10:07:44.921959 4815 generic.go:334] "Generic (PLEG): container finished" podID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerID="583e7e9a06709a466c71199e378754b4be29de55d07f6c2406cecf244fffc478" exitCode=0 Dec 05 10:07:44 crc kubenswrapper[4815]: I1205 10:07:44.922012 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgjw2" event={"ID":"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d","Type":"ContainerDied","Data":"583e7e9a06709a466c71199e378754b4be29de55d07f6c2406cecf244fffc478"} Dec 05 10:07:44 crc kubenswrapper[4815]: I1205 10:07:44.922063 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgjw2" event={"ID":"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d","Type":"ContainerStarted","Data":"68e8279da53695918163a9dcbcd17b7589d3e1120ae1ba7edad2cf78a8b4ada0"} Dec 05 10:07:45 crc kubenswrapper[4815]: I1205 10:07:45.932418 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgjw2" event={"ID":"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d","Type":"ContainerStarted","Data":"077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3"} Dec 05 10:07:50 crc kubenswrapper[4815]: I1205 10:07:50.192566 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:07:50 crc kubenswrapper[4815]: I1205 10:07:50.192928 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:07:51 crc kubenswrapper[4815]: I1205 10:07:51.059737 4815 generic.go:334] "Generic (PLEG): container finished" podID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerID="077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3" exitCode=0 Dec 05 10:07:51 crc kubenswrapper[4815]: I1205 10:07:51.059809 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgjw2" event={"ID":"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d","Type":"ContainerDied","Data":"077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3"} Dec 05 10:07:52 crc kubenswrapper[4815]: I1205 10:07:52.094994 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgjw2" event={"ID":"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d","Type":"ContainerStarted","Data":"6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd"} Dec 05 10:07:52 crc kubenswrapper[4815]: I1205 10:07:52.134097 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tgjw2" podStartSLOduration=3.464636148 podStartE2EDuration="10.134064328s" podCreationTimestamp="2025-12-05 10:07:42 +0000 UTC" firstStartedPulling="2025-12-05 10:07:44.92392824 +0000 UTC m=+3663.802535077" lastFinishedPulling="2025-12-05 10:07:51.5933564 +0000 UTC m=+3670.471963257" observedRunningTime="2025-12-05 10:07:52.122099981 +0000 UTC m=+3671.000706828" watchObservedRunningTime="2025-12-05 10:07:52.134064328 +0000 UTC m=+3671.012671165" Dec 05 10:07:53 crc kubenswrapper[4815]: I1205 10:07:53.106149 4815 generic.go:334] "Generic (PLEG): container finished" podID="e6418e66-6e27-4292-88c6-fca958ba3665" containerID="bbbed17c95d9aef3cfaa95ef744a4013fab3bac5e7bdfdf2be1676db2cac4a82" exitCode=0 Dec 05 10:07:53 crc kubenswrapper[4815]: I1205 10:07:53.106261 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" event={"ID":"e6418e66-6e27-4292-88c6-fca958ba3665","Type":"ContainerDied","Data":"bbbed17c95d9aef3cfaa95ef744a4013fab3bac5e7bdfdf2be1676db2cac4a82"} Dec 05 10:07:53 crc kubenswrapper[4815]: I1205 10:07:53.330290 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:53 crc kubenswrapper[4815]: I1205 10:07:53.330797 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.394196 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-tgjw2" podUID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerName="registry-server" probeResult="failure" output=< Dec 05 10:07:54 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 10:07:54 crc kubenswrapper[4815]: > Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.551705 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.750555 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-custom-ceph-combined-ca-bundle\") pod \"e6418e66-6e27-4292-88c6-fca958ba3665\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.750643 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ssh-key\") pod \"e6418e66-6e27-4292-88c6-fca958ba3665\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.750690 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2k5n\" (UniqueName: \"kubernetes.io/projected/e6418e66-6e27-4292-88c6-fca958ba3665-kube-api-access-d2k5n\") pod \"e6418e66-6e27-4292-88c6-fca958ba3665\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.750735 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-1\") pod \"e6418e66-6e27-4292-88c6-fca958ba3665\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.750778 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-nova-extra-config-0\") pod \"e6418e66-6e27-4292-88c6-fca958ba3665\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.750808 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-1\") pod \"e6418e66-6e27-4292-88c6-fca958ba3665\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.750855 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-0\") pod \"e6418e66-6e27-4292-88c6-fca958ba3665\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.750874 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-inventory\") pod \"e6418e66-6e27-4292-88c6-fca958ba3665\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.750896 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-0\") pod \"e6418e66-6e27-4292-88c6-fca958ba3665\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.750938 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-ceph-nova-0\") pod \"e6418e66-6e27-4292-88c6-fca958ba3665\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.750959 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ceph\") pod \"e6418e66-6e27-4292-88c6-fca958ba3665\" (UID: \"e6418e66-6e27-4292-88c6-fca958ba3665\") " Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.757857 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6418e66-6e27-4292-88c6-fca958ba3665-kube-api-access-d2k5n" (OuterVolumeSpecName: "kube-api-access-d2k5n") pod "e6418e66-6e27-4292-88c6-fca958ba3665" (UID: "e6418e66-6e27-4292-88c6-fca958ba3665"). InnerVolumeSpecName "kube-api-access-d2k5n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.758397 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "e6418e66-6e27-4292-88c6-fca958ba3665" (UID: "e6418e66-6e27-4292-88c6-fca958ba3665"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.777687 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ceph" (OuterVolumeSpecName: "ceph") pod "e6418e66-6e27-4292-88c6-fca958ba3665" (UID: "e6418e66-6e27-4292-88c6-fca958ba3665"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.792634 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "e6418e66-6e27-4292-88c6-fca958ba3665" (UID: "e6418e66-6e27-4292-88c6-fca958ba3665"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.796934 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-inventory" (OuterVolumeSpecName: "inventory") pod "e6418e66-6e27-4292-88c6-fca958ba3665" (UID: "e6418e66-6e27-4292-88c6-fca958ba3665"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.806695 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "e6418e66-6e27-4292-88c6-fca958ba3665" (UID: "e6418e66-6e27-4292-88c6-fca958ba3665"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.817731 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e6418e66-6e27-4292-88c6-fca958ba3665" (UID: "e6418e66-6e27-4292-88c6-fca958ba3665"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.833977 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "e6418e66-6e27-4292-88c6-fca958ba3665" (UID: "e6418e66-6e27-4292-88c6-fca958ba3665"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.840707 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "e6418e66-6e27-4292-88c6-fca958ba3665" (UID: "e6418e66-6e27-4292-88c6-fca958ba3665"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.853660 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.853692 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2k5n\" (UniqueName: \"kubernetes.io/projected/e6418e66-6e27-4292-88c6-fca958ba3665-kube-api-access-d2k5n\") on node \"crc\" DevicePath \"\"" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.853713 4815 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.853724 4815 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.853735 4815 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.853746 4815 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-inventory\") on node \"crc\" DevicePath \"\"" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.853756 4815 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/e6418e66-6e27-4292-88c6-fca958ba3665-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.853767 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.853805 4815 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.854331 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "e6418e66-6e27-4292-88c6-fca958ba3665" (UID: "e6418e66-6e27-4292-88c6-fca958ba3665"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.857958 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "e6418e66-6e27-4292-88c6-fca958ba3665" (UID: "e6418e66-6e27-4292-88c6-fca958ba3665"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.954877 4815 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 05 10:07:54 crc kubenswrapper[4815]: I1205 10:07:54.954914 4815 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/e6418e66-6e27-4292-88c6-fca958ba3665-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 05 10:07:55 crc kubenswrapper[4815]: I1205 10:07:55.124235 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" event={"ID":"e6418e66-6e27-4292-88c6-fca958ba3665","Type":"ContainerDied","Data":"3fb785a35dfd76520e84c917511a2ce376c9f64f1f8d521f1f75f181aad9c6d5"} Dec 05 10:07:55 crc kubenswrapper[4815]: I1205 10:07:55.124275 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3fb785a35dfd76520e84c917511a2ce376c9f64f1f8d521f1f75f181aad9c6d5" Dec 05 10:07:55 crc kubenswrapper[4815]: I1205 10:07:55.124334 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl" Dec 05 10:08:03 crc kubenswrapper[4815]: I1205 10:08:03.402811 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:08:03 crc kubenswrapper[4815]: I1205 10:08:03.472844 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:08:03 crc kubenswrapper[4815]: I1205 10:08:03.652168 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tgjw2"] Dec 05 10:08:05 crc kubenswrapper[4815]: I1205 10:08:05.230391 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tgjw2" podUID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerName="registry-server" containerID="cri-o://6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd" gracePeriod=2 Dec 05 10:08:05 crc kubenswrapper[4815]: I1205 10:08:05.713881 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:08:05 crc kubenswrapper[4815]: I1205 10:08:05.870625 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-catalog-content\") pod \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " Dec 05 10:08:05 crc kubenswrapper[4815]: I1205 10:08:05.870687 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj9th\" (UniqueName: \"kubernetes.io/projected/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-kube-api-access-gj9th\") pod \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " Dec 05 10:08:05 crc kubenswrapper[4815]: I1205 10:08:05.870778 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-utilities\") pod \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\" (UID: \"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d\") " Dec 05 10:08:05 crc kubenswrapper[4815]: I1205 10:08:05.872056 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-utilities" (OuterVolumeSpecName: "utilities") pod "a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" (UID: "a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:08:05 crc kubenswrapper[4815]: I1205 10:08:05.876976 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-kube-api-access-gj9th" (OuterVolumeSpecName: "kube-api-access-gj9th") pod "a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" (UID: "a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d"). InnerVolumeSpecName "kube-api-access-gj9th". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:08:05 crc kubenswrapper[4815]: I1205 10:08:05.972796 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj9th\" (UniqueName: \"kubernetes.io/projected/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-kube-api-access-gj9th\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:05 crc kubenswrapper[4815]: I1205 10:08:05.972831 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.020630 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" (UID: "a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.074759 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.244482 4815 generic.go:334] "Generic (PLEG): container finished" podID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerID="6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd" exitCode=0 Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.244624 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tgjw2" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.244612 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgjw2" event={"ID":"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d","Type":"ContainerDied","Data":"6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd"} Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.244825 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tgjw2" event={"ID":"a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d","Type":"ContainerDied","Data":"68e8279da53695918163a9dcbcd17b7589d3e1120ae1ba7edad2cf78a8b4ada0"} Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.244886 4815 scope.go:117] "RemoveContainer" containerID="6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.277941 4815 scope.go:117] "RemoveContainer" containerID="077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.323156 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tgjw2"] Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.331622 4815 scope.go:117] "RemoveContainer" containerID="583e7e9a06709a466c71199e378754b4be29de55d07f6c2406cecf244fffc478" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.337439 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tgjw2"] Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.355715 4815 scope.go:117] "RemoveContainer" containerID="6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd" Dec 05 10:08:06 crc kubenswrapper[4815]: E1205 10:08:06.356205 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd\": container with ID starting with 6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd not found: ID does not exist" containerID="6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.356242 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd"} err="failed to get container status \"6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd\": rpc error: code = NotFound desc = could not find container \"6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd\": container with ID starting with 6cfd777331741149eb25fec1334ebbce0d3387cf8767e4c19982461867c2b9fd not found: ID does not exist" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.356314 4815 scope.go:117] "RemoveContainer" containerID="077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3" Dec 05 10:08:06 crc kubenswrapper[4815]: E1205 10:08:06.356660 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3\": container with ID starting with 077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3 not found: ID does not exist" containerID="077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.356703 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3"} err="failed to get container status \"077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3\": rpc error: code = NotFound desc = could not find container \"077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3\": container with ID starting with 077e246402b2c9189a5fa90285a4344ae54b3c303b406a616e298950d80f22e3 not found: ID does not exist" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.356732 4815 scope.go:117] "RemoveContainer" containerID="583e7e9a06709a466c71199e378754b4be29de55d07f6c2406cecf244fffc478" Dec 05 10:08:06 crc kubenswrapper[4815]: E1205 10:08:06.357062 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"583e7e9a06709a466c71199e378754b4be29de55d07f6c2406cecf244fffc478\": container with ID starting with 583e7e9a06709a466c71199e378754b4be29de55d07f6c2406cecf244fffc478 not found: ID does not exist" containerID="583e7e9a06709a466c71199e378754b4be29de55d07f6c2406cecf244fffc478" Dec 05 10:08:06 crc kubenswrapper[4815]: I1205 10:08:06.357094 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"583e7e9a06709a466c71199e378754b4be29de55d07f6c2406cecf244fffc478"} err="failed to get container status \"583e7e9a06709a466c71199e378754b4be29de55d07f6c2406cecf244fffc478\": rpc error: code = NotFound desc = could not find container \"583e7e9a06709a466c71199e378754b4be29de55d07f6c2406cecf244fffc478\": container with ID starting with 583e7e9a06709a466c71199e378754b4be29de55d07f6c2406cecf244fffc478 not found: ID does not exist" Dec 05 10:08:07 crc kubenswrapper[4815]: I1205 10:08:07.427978 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" path="/var/lib/kubelet/pods/a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d/volumes" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.730723 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 05 10:08:11 crc kubenswrapper[4815]: E1205 10:08:11.731455 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerName="extract-content" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.731467 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerName="extract-content" Dec 05 10:08:11 crc kubenswrapper[4815]: E1205 10:08:11.731493 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerName="extract-utilities" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.731511 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerName="extract-utilities" Dec 05 10:08:11 crc kubenswrapper[4815]: E1205 10:08:11.731526 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6418e66-6e27-4292-88c6-fca958ba3665" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.731533 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6418e66-6e27-4292-88c6-fca958ba3665" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 05 10:08:11 crc kubenswrapper[4815]: E1205 10:08:11.731540 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerName="registry-server" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.731546 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerName="registry-server" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.731723 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6418e66-6e27-4292-88c6-fca958ba3665" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.731737 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7f594a3-2b8f-4e6c-a12d-4802ac9e2b4d" containerName="registry-server" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.732673 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.734477 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.738310 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.757873 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.817612 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.819383 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.822264 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.829713 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.891127 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.891184 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.891234 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.891269 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.891283 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.891824 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.892036 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-dev\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.892154 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-sys\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.892230 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.892300 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.892424 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-run\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.892566 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.892659 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.892778 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.892856 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzt7d\" (UniqueName: \"kubernetes.io/projected/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-kube-api-access-xzt7d\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.892963 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994285 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-run\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994343 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994359 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6f53fc9d-0e41-4083-855a-1624fb520e84-ceph\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994431 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-run\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994552 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-scripts\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994589 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994611 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994662 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994682 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994701 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzt7d\" (UniqueName: \"kubernetes.io/projected/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-kube-api-access-xzt7d\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994738 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994770 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994773 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994799 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-etc-nvme\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994817 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-sys\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994832 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994854 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994873 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994901 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994922 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgwtm\" (UniqueName: \"kubernetes.io/projected/6f53fc9d-0e41-4083-855a-1624fb520e84-kube-api-access-zgwtm\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994959 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.994975 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995001 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995016 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995040 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-run\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995091 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995121 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-config-data\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995137 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-dev\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995156 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-dev\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995174 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-sys\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995190 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-config-data-custom\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995210 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995230 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995262 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-lib-modules\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995639 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995718 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995774 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.995842 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.996710 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-dev\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.996786 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-sys\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.996828 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:11 crc kubenswrapper[4815]: I1205 10:08:11.996886 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.000342 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.001709 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.002172 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.002960 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.014723 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.015165 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzt7d\" (UniqueName: \"kubernetes.io/projected/15d9387d-aef1-48fe-8fb1-cb9d3dec22ea-kube-api-access-xzt7d\") pod \"cinder-volume-volume1-0\" (UID: \"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea\") " pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.055652 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096444 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-lib-modules\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096543 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096564 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6f53fc9d-0e41-4083-855a-1624fb520e84-ceph\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096596 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-scripts\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096632 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096681 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-etc-nvme\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096701 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096716 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-sys\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096758 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096782 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096803 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgwtm\" (UniqueName: \"kubernetes.io/projected/6f53fc9d-0e41-4083-855a-1624fb520e84-kube-api-access-zgwtm\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096829 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096851 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-run\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096892 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-dev\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096909 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-config-data\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.096937 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-config-data-custom\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.097346 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-etc-nvme\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.097338 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.097536 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.097672 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-sys\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.097797 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.097913 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-lib-modules\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.098033 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.098138 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.098870 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-run\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.098870 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6f53fc9d-0e41-4083-855a-1624fb520e84-dev\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.103268 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-scripts\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.103272 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6f53fc9d-0e41-4083-855a-1624fb520e84-ceph\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.105489 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-config-data\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.106037 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-config-data-custom\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.106871 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f53fc9d-0e41-4083-855a-1624fb520e84-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.120653 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgwtm\" (UniqueName: \"kubernetes.io/projected/6f53fc9d-0e41-4083-855a-1624fb520e84-kube-api-access-zgwtm\") pod \"cinder-backup-0\" (UID: \"6f53fc9d-0e41-4083-855a-1624fb520e84\") " pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.180432 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.401701 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-s8qsm"] Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.404543 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-s8qsm" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.410239 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-s8qsm"] Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.517418 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdb34d69-b26b-4258-9719-50a69ddce407-operator-scripts\") pod \"manila-db-create-s8qsm\" (UID: \"fdb34d69-b26b-4258-9719-50a69ddce407\") " pod="openstack/manila-db-create-s8qsm" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.517566 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwzf8\" (UniqueName: \"kubernetes.io/projected/fdb34d69-b26b-4258-9719-50a69ddce407-kube-api-access-wwzf8\") pod \"manila-db-create-s8qsm\" (UID: \"fdb34d69-b26b-4258-9719-50a69ddce407\") " pod="openstack/manila-db-create-s8qsm" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.621784 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdb34d69-b26b-4258-9719-50a69ddce407-operator-scripts\") pod \"manila-db-create-s8qsm\" (UID: \"fdb34d69-b26b-4258-9719-50a69ddce407\") " pod="openstack/manila-db-create-s8qsm" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.622146 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwzf8\" (UniqueName: \"kubernetes.io/projected/fdb34d69-b26b-4258-9719-50a69ddce407-kube-api-access-wwzf8\") pod \"manila-db-create-s8qsm\" (UID: \"fdb34d69-b26b-4258-9719-50a69ddce407\") " pod="openstack/manila-db-create-s8qsm" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.622176 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-1012-account-create-update-gd8rc"] Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.623665 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdb34d69-b26b-4258-9719-50a69ddce407-operator-scripts\") pod \"manila-db-create-s8qsm\" (UID: \"fdb34d69-b26b-4258-9719-50a69ddce407\") " pod="openstack/manila-db-create-s8qsm" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.623672 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1012-account-create-update-gd8rc" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.626598 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.649530 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.651151 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.661037 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwzf8\" (UniqueName: \"kubernetes.io/projected/fdb34d69-b26b-4258-9719-50a69ddce407-kube-api-access-wwzf8\") pod \"manila-db-create-s8qsm\" (UID: \"fdb34d69-b26b-4258-9719-50a69ddce407\") " pod="openstack/manila-db-create-s8qsm" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.661921 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.662665 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.662837 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-wng5w" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.662992 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.675582 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-1012-account-create-update-gd8rc"] Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.703575 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.725593 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7170277e-ef0a-4821-9c74-0c61204feabb-operator-scripts\") pod \"manila-1012-account-create-update-gd8rc\" (UID: \"7170277e-ef0a-4821-9c74-0c61204feabb\") " pod="openstack/manila-1012-account-create-update-gd8rc" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.725742 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb4h7\" (UniqueName: \"kubernetes.io/projected/7170277e-ef0a-4821-9c74-0c61204feabb-kube-api-access-fb4h7\") pod \"manila-1012-account-create-update-gd8rc\" (UID: \"7170277e-ef0a-4821-9c74-0c61204feabb\") " pod="openstack/manila-1012-account-create-update-gd8rc" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.783036 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-s8qsm" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.784219 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.823641 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.831678 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7rsl\" (UniqueName: \"kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-kube-api-access-x7rsl\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.833146 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.833275 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-scripts\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.833314 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.833661 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-ceph\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.833775 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7170277e-ef0a-4821-9c74-0c61204feabb-operator-scripts\") pod \"manila-1012-account-create-update-gd8rc\" (UID: \"7170277e-ef0a-4821-9c74-0c61204feabb\") " pod="openstack/manila-1012-account-create-update-gd8rc" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.833813 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-config-data\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.833146 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.833220 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.834735 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7170277e-ef0a-4821-9c74-0c61204feabb-operator-scripts\") pod \"manila-1012-account-create-update-gd8rc\" (UID: \"7170277e-ef0a-4821-9c74-0c61204feabb\") " pod="openstack/manila-1012-account-create-update-gd8rc" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.841834 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.841890 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.842164 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb4h7\" (UniqueName: \"kubernetes.io/projected/7170277e-ef0a-4821-9c74-0c61204feabb-kube-api-access-fb4h7\") pod \"manila-1012-account-create-update-gd8rc\" (UID: \"7170277e-ef0a-4821-9c74-0c61204feabb\") " pod="openstack/manila-1012-account-create-update-gd8rc" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.842198 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-logs\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.871315 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.896886 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb4h7\" (UniqueName: \"kubernetes.io/projected/7170277e-ef0a-4821-9c74-0c61204feabb-kube-api-access-fb4h7\") pod \"manila-1012-account-create-update-gd8rc\" (UID: \"7170277e-ef0a-4821-9c74-0c61204feabb\") " pod="openstack/manila-1012-account-create-update-gd8rc" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.901113 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6cd8dc9dcc-bvxhw"] Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.902745 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.913653 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.914038 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.914255 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.914467 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-r7jxd" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.966536 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.966606 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-scripts\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.966632 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.966652 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.966693 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-config-data\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.966724 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-ceph\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.966942 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-config-data\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.967005 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-logs\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.967458 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6cd8dc9dcc-bvxhw"] Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.967625 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-scripts\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.967653 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2vnk\" (UniqueName: \"kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-kube-api-access-j2vnk\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.967673 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.967689 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.967713 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.968112 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-ceph\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.968429 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.978991 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-config-data\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.988019 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-scripts\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.989087 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1012-account-create-update-gd8rc" Dec 05 10:08:12 crc kubenswrapper[4815]: I1205 10:08:12.989091 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-ceph\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:12.999777 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.006588 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.007120 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-logs\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.007165 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7rsl\" (UniqueName: \"kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-kube-api-access-x7rsl\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.007190 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.007629 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.008467 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-logs\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.017185 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.043545 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7rsl\" (UniqueName: \"kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-kube-api-access-x7rsl\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.047392 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:13 crc kubenswrapper[4815]: E1205 10:08:13.048349 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance httpd-run internal-tls-certs kube-api-access-j2vnk logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="29c22886-9562-4868-bde7-34faa5069c29" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.086076 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.110530 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-scripts\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.110727 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-scripts\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.110754 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2vnk\" (UniqueName: \"kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-kube-api-access-j2vnk\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.110806 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.110822 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-ceph\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.111073 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.111096 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-horizon-secret-key\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.111127 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-config-data\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.116080 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-scripts\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.116326 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.122244 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.126059 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.128153 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-logs\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.128323 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sc8n5\" (UniqueName: \"kubernetes.io/projected/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-kube-api-access-sc8n5\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.128790 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.131335 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.131447 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-config-data\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.131591 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-logs\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.132147 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-logs\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.133973 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.146637 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-ceph\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.147217 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2vnk\" (UniqueName: \"kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-kube-api-access-j2vnk\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.158390 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-config-data\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.172130 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.188633 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-cbc799c5c-zhk6n"] Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.192788 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.225099 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.237772 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cbc799c5c-zhk6n"] Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.240253 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-horizon-secret-key\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.240306 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-config-data\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.249803 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-config-data\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.249935 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-logs\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.250014 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sc8n5\" (UniqueName: \"kubernetes.io/projected/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-kube-api-access-sc8n5\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.250335 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-config-data\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.250595 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/92a9f228-e917-4610-bc27-822d3e0e8578-horizon-secret-key\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.250707 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-scripts\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.250728 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92a9f228-e917-4610-bc27-822d3e0e8578-logs\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.250835 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-scripts\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.250870 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsqhf\" (UniqueName: \"kubernetes.io/projected/92a9f228-e917-4610-bc27-822d3e0e8578-kube-api-access-xsqhf\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.251237 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-horizon-secret-key\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.251258 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-logs\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.251911 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.252639 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.262889 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-scripts\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.266425 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.278342 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sc8n5\" (UniqueName: \"kubernetes.io/projected/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-kube-api-access-sc8n5\") pod \"horizon-6cd8dc9dcc-bvxhw\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.301006 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.312893 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea","Type":"ContainerStarted","Data":"0e93565044f7a07b2d938bd20bb35bde1cd15dc8444435d6ceca841815493e9d"} Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.314471 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.315295 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"6f53fc9d-0e41-4083-855a-1624fb520e84","Type":"ContainerStarted","Data":"482a1be2b38c49be7fa6a4e7e37e30e52cd675390c0ba7dc19bd06de70d32fd9"} Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.344306 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.352436 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-config-data\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.352559 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/92a9f228-e917-4610-bc27-822d3e0e8578-horizon-secret-key\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.352616 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-scripts\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.352638 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92a9f228-e917-4610-bc27-822d3e0e8578-logs\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.352691 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsqhf\" (UniqueName: \"kubernetes.io/projected/92a9f228-e917-4610-bc27-822d3e0e8578-kube-api-access-xsqhf\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.354761 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-scripts\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.354934 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92a9f228-e917-4610-bc27-822d3e0e8578-logs\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.355273 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-config-data\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.360748 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/92a9f228-e917-4610-bc27-822d3e0e8578-horizon-secret-key\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.399455 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsqhf\" (UniqueName: \"kubernetes.io/projected/92a9f228-e917-4610-bc27-822d3e0e8578-kube-api-access-xsqhf\") pod \"horizon-cbc799c5c-zhk6n\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.453960 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-ceph\") pod \"29c22886-9562-4868-bde7-34faa5069c29\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.454047 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2vnk\" (UniqueName: \"kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-kube-api-access-j2vnk\") pod \"29c22886-9562-4868-bde7-34faa5069c29\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.454071 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-logs\") pod \"29c22886-9562-4868-bde7-34faa5069c29\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.454103 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-config-data\") pod \"29c22886-9562-4868-bde7-34faa5069c29\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.454187 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-internal-tls-certs\") pod \"29c22886-9562-4868-bde7-34faa5069c29\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.454237 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-combined-ca-bundle\") pod \"29c22886-9562-4868-bde7-34faa5069c29\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.454270 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-httpd-run\") pod \"29c22886-9562-4868-bde7-34faa5069c29\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.454349 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-scripts\") pod \"29c22886-9562-4868-bde7-34faa5069c29\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.454381 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"29c22886-9562-4868-bde7-34faa5069c29\" (UID: \"29c22886-9562-4868-bde7-34faa5069c29\") " Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.457068 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "29c22886-9562-4868-bde7-34faa5069c29" (UID: "29c22886-9562-4868-bde7-34faa5069c29"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.458642 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-logs" (OuterVolumeSpecName: "logs") pod "29c22886-9562-4868-bde7-34faa5069c29" (UID: "29c22886-9562-4868-bde7-34faa5069c29"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.460854 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "29c22886-9562-4868-bde7-34faa5069c29" (UID: "29c22886-9562-4868-bde7-34faa5069c29"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.462644 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-kube-api-access-j2vnk" (OuterVolumeSpecName: "kube-api-access-j2vnk") pod "29c22886-9562-4868-bde7-34faa5069c29" (UID: "29c22886-9562-4868-bde7-34faa5069c29"). InnerVolumeSpecName "kube-api-access-j2vnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.462794 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-ceph" (OuterVolumeSpecName: "ceph") pod "29c22886-9562-4868-bde7-34faa5069c29" (UID: "29c22886-9562-4868-bde7-34faa5069c29"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.462853 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "29c22886-9562-4868-bde7-34faa5069c29" (UID: "29c22886-9562-4868-bde7-34faa5069c29"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.463459 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-scripts" (OuterVolumeSpecName: "scripts") pod "29c22886-9562-4868-bde7-34faa5069c29" (UID: "29c22886-9562-4868-bde7-34faa5069c29"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.475930 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-config-data" (OuterVolumeSpecName: "config-data") pod "29c22886-9562-4868-bde7-34faa5069c29" (UID: "29c22886-9562-4868-bde7-34faa5069c29"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.477415 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "29c22886-9562-4868-bde7-34faa5069c29" (UID: "29c22886-9562-4868-bde7-34faa5069c29"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.508279 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-s8qsm"] Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.563391 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.564985 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.565034 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.565050 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.565062 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2vnk\" (UniqueName: \"kubernetes.io/projected/29c22886-9562-4868-bde7-34faa5069c29-kube-api-access-j2vnk\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.565074 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-logs\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.565085 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.565096 4815 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.565110 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/29c22886-9562-4868-bde7-34faa5069c29-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.565121 4815 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/29c22886-9562-4868-bde7-34faa5069c29-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.596770 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.667629 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:13 crc kubenswrapper[4815]: I1205 10:08:13.780193 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-1012-account-create-update-gd8rc"] Dec 05 10:08:13 crc kubenswrapper[4815]: W1205 10:08:13.796082 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7170277e_ef0a_4821_9c74_0c61204feabb.slice/crio-35d54285fbae557fcee93ff7301798f95081d90592cdda351691f3ac72b6a796 WatchSource:0}: Error finding container 35d54285fbae557fcee93ff7301798f95081d90592cdda351691f3ac72b6a796: Status 404 returned error can't find the container with id 35d54285fbae557fcee93ff7301798f95081d90592cdda351691f3ac72b6a796 Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.011540 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6cd8dc9dcc-bvxhw"] Dec 05 10:08:14 crc kubenswrapper[4815]: W1205 10:08:14.032171 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf69fb20_ecce_4a27_a4c1_5ff38416c9e1.slice/crio-5c63fdc252f9135984102cac80fddb6c5684fab2c7fb550cd2a08de7282e78d9 WatchSource:0}: Error finding container 5c63fdc252f9135984102cac80fddb6c5684fab2c7fb550cd2a08de7282e78d9: Status 404 returned error can't find the container with id 5c63fdc252f9135984102cac80fddb6c5684fab2c7fb550cd2a08de7282e78d9 Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.205287 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.258475 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-cbc799c5c-zhk6n"] Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.354720 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-1012-account-create-update-gd8rc" event={"ID":"7170277e-ef0a-4821-9c74-0c61204feabb","Type":"ContainerStarted","Data":"57fc7ec8e61b912927d3065c9dce1ced121427a9daa73cb6b5e38572849004fc"} Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.354767 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-1012-account-create-update-gd8rc" event={"ID":"7170277e-ef0a-4821-9c74-0c61204feabb","Type":"ContainerStarted","Data":"35d54285fbae557fcee93ff7301798f95081d90592cdda351691f3ac72b6a796"} Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.380274 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e5ff121f-57a1-4f4a-85b4-691b465eef1c","Type":"ContainerStarted","Data":"7806a34a94119bf10bf9326fa370231bda6a3e2a1e1c9162461843bbb06d63be"} Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.387932 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-1012-account-create-update-gd8rc" podStartSLOduration=2.387912954 podStartE2EDuration="2.387912954s" podCreationTimestamp="2025-12-05 10:08:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:08:14.380136452 +0000 UTC m=+3693.258743299" watchObservedRunningTime="2025-12-05 10:08:14.387912954 +0000 UTC m=+3693.266519791" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.405721 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-s8qsm" event={"ID":"fdb34d69-b26b-4258-9719-50a69ddce407","Type":"ContainerStarted","Data":"f7b9227bdcbb51659cdcb9e31207a69936a648a3c9df4cb86dc8f35e403e7627"} Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.405770 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-s8qsm" event={"ID":"fdb34d69-b26b-4258-9719-50a69ddce407","Type":"ContainerStarted","Data":"091f1a52f2fe0e88c4ebbc47abbbbb5ed30bbb59efc6ee375f2cd4b6835a0979"} Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.409817 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cd8dc9dcc-bvxhw" event={"ID":"af69fb20-ecce-4a27-a4c1-5ff38416c9e1","Type":"ContainerStarted","Data":"5c63fdc252f9135984102cac80fddb6c5684fab2c7fb550cd2a08de7282e78d9"} Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.411286 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.419677 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cbc799c5c-zhk6n" event={"ID":"92a9f228-e917-4610-bc27-822d3e0e8578","Type":"ContainerStarted","Data":"9fd67bce401ba820ad1cae4a96dfe211cba0e5902cc5fd1546f6691205f9c671"} Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.437463 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-create-s8qsm" podStartSLOduration=2.437440915 podStartE2EDuration="2.437440915s" podCreationTimestamp="2025-12-05 10:08:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:08:14.431145693 +0000 UTC m=+3693.309752530" watchObservedRunningTime="2025-12-05 10:08:14.437440915 +0000 UTC m=+3693.316047752" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.635034 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.660595 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.692558 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.694192 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.703124 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.703382 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.717279 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.830456 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.830528 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-logs\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.830557 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.830624 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.830681 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsfbs\" (UniqueName: \"kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-kube-api-access-nsfbs\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.830731 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.830754 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.830779 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.830797 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.932344 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.934187 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.934228 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.934256 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.934338 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.934364 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-logs\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.934391 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.934469 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.934566 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsfbs\" (UniqueName: \"kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-kube-api-access-nsfbs\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.934650 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.934816 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.935858 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-logs\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.938814 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.942165 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-ceph\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.945090 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.946446 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.946939 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.953648 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsfbs\" (UniqueName: \"kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-kube-api-access-nsfbs\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:14 crc kubenswrapper[4815]: I1205 10:08:14.976581 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:15 crc kubenswrapper[4815]: I1205 10:08:15.028688 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:15 crc kubenswrapper[4815]: I1205 10:08:15.429947 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="29c22886-9562-4868-bde7-34faa5069c29" path="/var/lib/kubelet/pods/29c22886-9562-4868-bde7-34faa5069c29/volumes" Dec 05 10:08:15 crc kubenswrapper[4815]: I1205 10:08:15.436447 4815 generic.go:334] "Generic (PLEG): container finished" podID="7170277e-ef0a-4821-9c74-0c61204feabb" containerID="57fc7ec8e61b912927d3065c9dce1ced121427a9daa73cb6b5e38572849004fc" exitCode=0 Dec 05 10:08:15 crc kubenswrapper[4815]: I1205 10:08:15.436566 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-1012-account-create-update-gd8rc" event={"ID":"7170277e-ef0a-4821-9c74-0c61204feabb","Type":"ContainerDied","Data":"57fc7ec8e61b912927d3065c9dce1ced121427a9daa73cb6b5e38572849004fc"} Dec 05 10:08:15 crc kubenswrapper[4815]: I1205 10:08:15.439344 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e5ff121f-57a1-4f4a-85b4-691b465eef1c","Type":"ContainerStarted","Data":"00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f"} Dec 05 10:08:15 crc kubenswrapper[4815]: I1205 10:08:15.440816 4815 generic.go:334] "Generic (PLEG): container finished" podID="fdb34d69-b26b-4258-9719-50a69ddce407" containerID="f7b9227bdcbb51659cdcb9e31207a69936a648a3c9df4cb86dc8f35e403e7627" exitCode=0 Dec 05 10:08:15 crc kubenswrapper[4815]: I1205 10:08:15.440842 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-s8qsm" event={"ID":"fdb34d69-b26b-4258-9719-50a69ddce407","Type":"ContainerDied","Data":"f7b9227bdcbb51659cdcb9e31207a69936a648a3c9df4cb86dc8f35e403e7627"} Dec 05 10:08:15 crc kubenswrapper[4815]: I1205 10:08:15.831155 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:15 crc kubenswrapper[4815]: W1205 10:08:15.941840 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee526317_f5a1_4682_b2ab_612ce15d9bf9.slice/crio-125ea7302ad699d89f039c0b0631723a8f44d8efc69516d2ae3964113cef555c WatchSource:0}: Error finding container 125ea7302ad699d89f039c0b0631723a8f44d8efc69516d2ae3964113cef555c: Status 404 returned error can't find the container with id 125ea7302ad699d89f039c0b0631723a8f44d8efc69516d2ae3964113cef555c Dec 05 10:08:15 crc kubenswrapper[4815]: I1205 10:08:15.967283 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6cd8dc9dcc-bvxhw"] Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.027632 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-c444bf898-b6z6t"] Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.030177 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.035530 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.144855 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.169633 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-tls-certs\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.169699 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-secret-key\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.169733 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-config-data\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.169780 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m2lg6\" (UniqueName: \"kubernetes.io/projected/819b83db-6f8e-4b8d-81d7-719760d1ff6b-kube-api-access-m2lg6\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.169812 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-combined-ca-bundle\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.169840 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/819b83db-6f8e-4b8d-81d7-719760d1ff6b-logs\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.169890 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-scripts\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.195333 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c444bf898-b6z6t"] Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.247197 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-cbc799c5c-zhk6n"] Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.280786 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m2lg6\" (UniqueName: \"kubernetes.io/projected/819b83db-6f8e-4b8d-81d7-719760d1ff6b-kube-api-access-m2lg6\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.280862 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-combined-ca-bundle\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.280898 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/819b83db-6f8e-4b8d-81d7-719760d1ff6b-logs\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.280966 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-scripts\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.281011 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-tls-certs\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.281066 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-secret-key\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.281101 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-config-data\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.282231 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-scripts\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.282353 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-config-data\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.284434 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/819b83db-6f8e-4b8d-81d7-719760d1ff6b-logs\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.304594 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-combined-ca-bundle\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.305361 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-tls-certs\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.320872 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-74b8f449c4-9gwqr"] Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.322631 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.358247 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74b8f449c4-9gwqr"] Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.429554 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-secret-key\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.435294 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m2lg6\" (UniqueName: \"kubernetes.io/projected/819b83db-6f8e-4b8d-81d7-719760d1ff6b-kube-api-access-m2lg6\") pod \"horizon-c444bf898-b6z6t\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.473581 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea","Type":"ContainerStarted","Data":"dca204645c9a28b1cac17638aa206c7d5dfab98c841b43b4fd07dc9eb9261b99"} Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.473839 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"15d9387d-aef1-48fe-8fb1-cb9d3dec22ea","Type":"ContainerStarted","Data":"aa8d269062bfe144b3af62adb1135e6c15b5b95476bf73fd3141736a53ad019d"} Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.476208 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"6f53fc9d-0e41-4083-855a-1624fb520e84","Type":"ContainerStarted","Data":"03391b538ade714638c6023aefdc929a4a80d4e90e854f0d7bc981e54b9bbd08"} Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.481156 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee526317-f5a1-4682-b2ab-612ce15d9bf9","Type":"ContainerStarted","Data":"125ea7302ad699d89f039c0b0631723a8f44d8efc69516d2ae3964113cef555c"} Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.483568 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.510559 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stz8x\" (UniqueName: \"kubernetes.io/projected/e9604337-767f-492a-aa51-f31578f44ece-kube-api-access-stz8x\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.510647 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9604337-767f-492a-aa51-f31578f44ece-logs\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.510699 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9604337-767f-492a-aa51-f31578f44ece-scripts\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.510800 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e9604337-767f-492a-aa51-f31578f44ece-config-data\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.510834 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e9604337-767f-492a-aa51-f31578f44ece-horizon-secret-key\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.510861 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9604337-767f-492a-aa51-f31578f44ece-horizon-tls-certs\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.510897 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9604337-767f-492a-aa51-f31578f44ece-combined-ca-bundle\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.536807 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=3.382377333 podStartE2EDuration="5.536784749s" podCreationTimestamp="2025-12-05 10:08:11 +0000 UTC" firstStartedPulling="2025-12-05 10:08:12.98023544 +0000 UTC m=+3691.858842277" lastFinishedPulling="2025-12-05 10:08:15.134642856 +0000 UTC m=+3694.013249693" observedRunningTime="2025-12-05 10:08:16.531372641 +0000 UTC m=+3695.409979478" watchObservedRunningTime="2025-12-05 10:08:16.536784749 +0000 UTC m=+3695.415391586" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.613540 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stz8x\" (UniqueName: \"kubernetes.io/projected/e9604337-767f-492a-aa51-f31578f44ece-kube-api-access-stz8x\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.613683 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9604337-767f-492a-aa51-f31578f44ece-logs\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.613767 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9604337-767f-492a-aa51-f31578f44ece-scripts\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.613942 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e9604337-767f-492a-aa51-f31578f44ece-config-data\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.614008 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e9604337-767f-492a-aa51-f31578f44ece-horizon-secret-key\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.614083 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9604337-767f-492a-aa51-f31578f44ece-horizon-tls-certs\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.614118 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9604337-767f-492a-aa51-f31578f44ece-combined-ca-bundle\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.618188 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e9604337-767f-492a-aa51-f31578f44ece-logs\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.618883 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/e9604337-767f-492a-aa51-f31578f44ece-config-data\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.619047 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e9604337-767f-492a-aa51-f31578f44ece-scripts\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.629897 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9604337-767f-492a-aa51-f31578f44ece-combined-ca-bundle\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.631838 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/e9604337-767f-492a-aa51-f31578f44ece-horizon-tls-certs\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.636368 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stz8x\" (UniqueName: \"kubernetes.io/projected/e9604337-767f-492a-aa51-f31578f44ece-kube-api-access-stz8x\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.652557 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/e9604337-767f-492a-aa51-f31578f44ece-horizon-secret-key\") pod \"horizon-74b8f449c4-9gwqr\" (UID: \"e9604337-767f-492a-aa51-f31578f44ece\") " pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:16 crc kubenswrapper[4815]: I1205 10:08:16.900892 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.061869 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.408290 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1012-account-create-update-gd8rc" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.425821 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-s8qsm" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.512025 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"6f53fc9d-0e41-4083-855a-1624fb520e84","Type":"ContainerStarted","Data":"59473cbf6f2b6a79da8df5e93e5f055d90247fa65a88f0c54aeca4fd2e62093a"} Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.533113 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e5ff121f-57a1-4f4a-85b4-691b465eef1c","Type":"ContainerStarted","Data":"91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235"} Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.533271 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e5ff121f-57a1-4f4a-85b4-691b465eef1c" containerName="glance-log" containerID="cri-o://00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f" gracePeriod=30 Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.533551 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="e5ff121f-57a1-4f4a-85b4-691b465eef1c" containerName="glance-httpd" containerID="cri-o://91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235" gracePeriod=30 Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.534866 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-c444bf898-b6z6t"] Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.540756 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-s8qsm" event={"ID":"fdb34d69-b26b-4258-9719-50a69ddce407","Type":"ContainerDied","Data":"091f1a52f2fe0e88c4ebbc47abbbbb5ed30bbb59efc6ee375f2cd4b6835a0979"} Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.540793 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="091f1a52f2fe0e88c4ebbc47abbbbb5ed30bbb59efc6ee375f2cd4b6835a0979" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.540861 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-s8qsm" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.551421 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwzf8\" (UniqueName: \"kubernetes.io/projected/fdb34d69-b26b-4258-9719-50a69ddce407-kube-api-access-wwzf8\") pod \"fdb34d69-b26b-4258-9719-50a69ddce407\" (UID: \"fdb34d69-b26b-4258-9719-50a69ddce407\") " Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.551610 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdb34d69-b26b-4258-9719-50a69ddce407-operator-scripts\") pod \"fdb34d69-b26b-4258-9719-50a69ddce407\" (UID: \"fdb34d69-b26b-4258-9719-50a69ddce407\") " Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.551683 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7170277e-ef0a-4821-9c74-0c61204feabb-operator-scripts\") pod \"7170277e-ef0a-4821-9c74-0c61204feabb\" (UID: \"7170277e-ef0a-4821-9c74-0c61204feabb\") " Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.551790 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fb4h7\" (UniqueName: \"kubernetes.io/projected/7170277e-ef0a-4821-9c74-0c61204feabb-kube-api-access-fb4h7\") pod \"7170277e-ef0a-4821-9c74-0c61204feabb\" (UID: \"7170277e-ef0a-4821-9c74-0c61204feabb\") " Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.554569 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fdb34d69-b26b-4258-9719-50a69ddce407-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fdb34d69-b26b-4258-9719-50a69ddce407" (UID: "fdb34d69-b26b-4258-9719-50a69ddce407"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.554931 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7170277e-ef0a-4821-9c74-0c61204feabb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7170277e-ef0a-4821-9c74-0c61204feabb" (UID: "7170277e-ef0a-4821-9c74-0c61204feabb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.564479 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-1012-account-create-update-gd8rc" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.564663 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-1012-account-create-update-gd8rc" event={"ID":"7170277e-ef0a-4821-9c74-0c61204feabb","Type":"ContainerDied","Data":"35d54285fbae557fcee93ff7301798f95081d90592cdda351691f3ac72b6a796"} Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.564685 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35d54285fbae557fcee93ff7301798f95081d90592cdda351691f3ac72b6a796" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.576120 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=4.106914569 podStartE2EDuration="6.576101313s" podCreationTimestamp="2025-12-05 10:08:11 +0000 UTC" firstStartedPulling="2025-12-05 10:08:13.176604667 +0000 UTC m=+3692.055211504" lastFinishedPulling="2025-12-05 10:08:15.645791411 +0000 UTC m=+3694.524398248" observedRunningTime="2025-12-05 10:08:17.544960684 +0000 UTC m=+3696.423567521" watchObservedRunningTime="2025-12-05 10:08:17.576101313 +0000 UTC m=+3696.454708150" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.582696 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7170277e-ef0a-4821-9c74-0c61204feabb-kube-api-access-fb4h7" (OuterVolumeSpecName: "kube-api-access-fb4h7") pod "7170277e-ef0a-4821-9c74-0c61204feabb" (UID: "7170277e-ef0a-4821-9c74-0c61204feabb"). InnerVolumeSpecName "kube-api-access-fb4h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.587045 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdb34d69-b26b-4258-9719-50a69ddce407-kube-api-access-wwzf8" (OuterVolumeSpecName: "kube-api-access-wwzf8") pod "fdb34d69-b26b-4258-9719-50a69ddce407" (UID: "fdb34d69-b26b-4258-9719-50a69ddce407"). InnerVolumeSpecName "kube-api-access-wwzf8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.605182 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.605161896 podStartE2EDuration="6.605161896s" podCreationTimestamp="2025-12-05 10:08:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:08:17.585703005 +0000 UTC m=+3696.464309842" watchObservedRunningTime="2025-12-05 10:08:17.605161896 +0000 UTC m=+3696.483768733" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.655327 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fdb34d69-b26b-4258-9719-50a69ddce407-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.655368 4815 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7170277e-ef0a-4821-9c74-0c61204feabb-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.655383 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fb4h7\" (UniqueName: \"kubernetes.io/projected/7170277e-ef0a-4821-9c74-0c61204feabb-kube-api-access-fb4h7\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.655395 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwzf8\" (UniqueName: \"kubernetes.io/projected/fdb34d69-b26b-4258-9719-50a69ddce407-kube-api-access-wwzf8\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:17 crc kubenswrapper[4815]: I1205 10:08:17.858198 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-74b8f449c4-9gwqr"] Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.342709 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.484212 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7rsl\" (UniqueName: \"kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-kube-api-access-x7rsl\") pod \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.484541 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-public-tls-certs\") pod \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.485063 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-combined-ca-bundle\") pod \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.485181 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-httpd-run\") pod \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.485225 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.485293 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-config-data\") pod \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.485381 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-scripts\") pod \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.485447 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-ceph\") pod \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.485473 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-logs\") pod \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\" (UID: \"e5ff121f-57a1-4f4a-85b4-691b465eef1c\") " Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.486540 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e5ff121f-57a1-4f4a-85b4-691b465eef1c" (UID: "e5ff121f-57a1-4f4a-85b4-691b465eef1c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.487542 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-logs" (OuterVolumeSpecName: "logs") pod "e5ff121f-57a1-4f4a-85b4-691b465eef1c" (UID: "e5ff121f-57a1-4f4a-85b4-691b465eef1c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.488473 4815 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.488535 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5ff121f-57a1-4f4a-85b4-691b465eef1c-logs\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.504551 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-kube-api-access-x7rsl" (OuterVolumeSpecName: "kube-api-access-x7rsl") pod "e5ff121f-57a1-4f4a-85b4-691b465eef1c" (UID: "e5ff121f-57a1-4f4a-85b4-691b465eef1c"). InnerVolumeSpecName "kube-api-access-x7rsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.509332 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-scripts" (OuterVolumeSpecName: "scripts") pod "e5ff121f-57a1-4f4a-85b4-691b465eef1c" (UID: "e5ff121f-57a1-4f4a-85b4-691b465eef1c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.536650 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "e5ff121f-57a1-4f4a-85b4-691b465eef1c" (UID: "e5ff121f-57a1-4f4a-85b4-691b465eef1c"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.554773 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-ceph" (OuterVolumeSpecName: "ceph") pod "e5ff121f-57a1-4f4a-85b4-691b465eef1c" (UID: "e5ff121f-57a1-4f4a-85b4-691b465eef1c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.592828 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.592873 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.592885 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.592897 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7rsl\" (UniqueName: \"kubernetes.io/projected/e5ff121f-57a1-4f4a-85b4-691b465eef1c-kube-api-access-x7rsl\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.604904 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74b8f449c4-9gwqr" event={"ID":"e9604337-767f-492a-aa51-f31578f44ece","Type":"ContainerStarted","Data":"33de378e5725340e93340863a639bc847710dc9160866ae9eb1b15ce836dc8dd"} Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.623158 4815 generic.go:334] "Generic (PLEG): container finished" podID="e5ff121f-57a1-4f4a-85b4-691b465eef1c" containerID="91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235" exitCode=143 Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.623458 4815 generic.go:334] "Generic (PLEG): container finished" podID="e5ff121f-57a1-4f4a-85b4-691b465eef1c" containerID="00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f" exitCode=143 Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.624011 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.624166 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e5ff121f-57a1-4f4a-85b4-691b465eef1c","Type":"ContainerDied","Data":"91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235"} Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.624240 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e5ff121f-57a1-4f4a-85b4-691b465eef1c","Type":"ContainerDied","Data":"00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f"} Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.624251 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"e5ff121f-57a1-4f4a-85b4-691b465eef1c","Type":"ContainerDied","Data":"7806a34a94119bf10bf9326fa370231bda6a3e2a1e1c9162461843bbb06d63be"} Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.624266 4815 scope.go:117] "RemoveContainer" containerID="91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.636042 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e5ff121f-57a1-4f4a-85b4-691b465eef1c" (UID: "e5ff121f-57a1-4f4a-85b4-691b465eef1c"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.640391 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.646398 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee526317-f5a1-4682-b2ab-612ce15d9bf9","Type":"ContainerStarted","Data":"a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510"} Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.654350 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c444bf898-b6z6t" event={"ID":"819b83db-6f8e-4b8d-81d7-719760d1ff6b","Type":"ContainerStarted","Data":"928391e4b689d1a0e503c4b5e20313b29717362c6d33205338f6b72cfdbf0292"} Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.664151 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-config-data" (OuterVolumeSpecName: "config-data") pod "e5ff121f-57a1-4f4a-85b4-691b465eef1c" (UID: "e5ff121f-57a1-4f4a-85b4-691b465eef1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.674210 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5ff121f-57a1-4f4a-85b4-691b465eef1c" (UID: "e5ff121f-57a1-4f4a-85b4-691b465eef1c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.694893 4815 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.694922 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.694932 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.694942 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5ff121f-57a1-4f4a-85b4-691b465eef1c-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.712623 4815 scope.go:117] "RemoveContainer" containerID="00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.764832 4815 scope.go:117] "RemoveContainer" containerID="91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235" Dec 05 10:08:18 crc kubenswrapper[4815]: E1205 10:08:18.768322 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235\": container with ID starting with 91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235 not found: ID does not exist" containerID="91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.768390 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235"} err="failed to get container status \"91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235\": rpc error: code = NotFound desc = could not find container \"91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235\": container with ID starting with 91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235 not found: ID does not exist" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.768425 4815 scope.go:117] "RemoveContainer" containerID="00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f" Dec 05 10:08:18 crc kubenswrapper[4815]: E1205 10:08:18.769167 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f\": container with ID starting with 00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f not found: ID does not exist" containerID="00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.769199 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f"} err="failed to get container status \"00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f\": rpc error: code = NotFound desc = could not find container \"00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f\": container with ID starting with 00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f not found: ID does not exist" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.769218 4815 scope.go:117] "RemoveContainer" containerID="91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.769548 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235"} err="failed to get container status \"91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235\": rpc error: code = NotFound desc = could not find container \"91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235\": container with ID starting with 91d3fe899572b5470c15bb5c759bf284a75dfb7722da0eca28e620a82328d235 not found: ID does not exist" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.769599 4815 scope.go:117] "RemoveContainer" containerID="00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.769995 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f"} err="failed to get container status \"00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f\": rpc error: code = NotFound desc = could not find container \"00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f\": container with ID starting with 00baf6dd0f8c18b95ff3b5482c3b2a4b25cafb960fc2f4ad63c73103d31e597f not found: ID does not exist" Dec 05 10:08:18 crc kubenswrapper[4815]: I1205 10:08:18.987155 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.001457 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.023850 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 10:08:19 crc kubenswrapper[4815]: E1205 10:08:19.024288 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb34d69-b26b-4258-9719-50a69ddce407" containerName="mariadb-database-create" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.024304 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb34d69-b26b-4258-9719-50a69ddce407" containerName="mariadb-database-create" Dec 05 10:08:19 crc kubenswrapper[4815]: E1205 10:08:19.024339 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5ff121f-57a1-4f4a-85b4-691b465eef1c" containerName="glance-httpd" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.024348 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5ff121f-57a1-4f4a-85b4-691b465eef1c" containerName="glance-httpd" Dec 05 10:08:19 crc kubenswrapper[4815]: E1205 10:08:19.024362 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7170277e-ef0a-4821-9c74-0c61204feabb" containerName="mariadb-account-create-update" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.024368 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7170277e-ef0a-4821-9c74-0c61204feabb" containerName="mariadb-account-create-update" Dec 05 10:08:19 crc kubenswrapper[4815]: E1205 10:08:19.024385 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5ff121f-57a1-4f4a-85b4-691b465eef1c" containerName="glance-log" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.024390 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5ff121f-57a1-4f4a-85b4-691b465eef1c" containerName="glance-log" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.024610 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5ff121f-57a1-4f4a-85b4-691b465eef1c" containerName="glance-log" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.024640 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdb34d69-b26b-4258-9719-50a69ddce407" containerName="mariadb-database-create" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.024649 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5ff121f-57a1-4f4a-85b4-691b465eef1c" containerName="glance-httpd" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.024662 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7170277e-ef0a-4821-9c74-0c61204feabb" containerName="mariadb-account-create-update" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.025734 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.025756 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.030522 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.030707 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.107007 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.107082 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm9qv\" (UniqueName: \"kubernetes.io/projected/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-kube-api-access-xm9qv\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.107439 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-config-data\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.107605 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-scripts\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.107650 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-ceph\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.107684 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.107947 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.107991 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.108110 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-logs\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.209626 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.209677 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.209722 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-logs\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.209759 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.209782 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm9qv\" (UniqueName: \"kubernetes.io/projected/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-kube-api-access-xm9qv\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.209812 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-config-data\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.209830 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-scripts\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.209847 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-ceph\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.209861 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.210015 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.210354 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-logs\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.210582 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.239577 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.239999 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-scripts\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.240443 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.240852 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-ceph\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.241762 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-config-data\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.259820 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.262119 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm9qv\" (UniqueName: \"kubernetes.io/projected/2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b-kube-api-access-xm9qv\") pod \"glance-default-external-api-0\" (UID: \"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b\") " pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.391429 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.464276 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5ff121f-57a1-4f4a-85b4-691b465eef1c" path="/var/lib/kubelet/pods/e5ff121f-57a1-4f4a-85b4-691b465eef1c/volumes" Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.690599 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee526317-f5a1-4682-b2ab-612ce15d9bf9","Type":"ContainerStarted","Data":"3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39"} Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.690813 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ee526317-f5a1-4682-b2ab-612ce15d9bf9" containerName="glance-log" containerID="cri-o://a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510" gracePeriod=30 Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.690918 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="ee526317-f5a1-4682-b2ab-612ce15d9bf9" containerName="glance-httpd" containerID="cri-o://3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39" gracePeriod=30 Dec 05 10:08:19 crc kubenswrapper[4815]: I1205 10:08:19.719405 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.719386195 podStartE2EDuration="5.719386195s" podCreationTimestamp="2025-12-05 10:08:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:08:19.707643375 +0000 UTC m=+3698.586250212" watchObservedRunningTime="2025-12-05 10:08:19.719386195 +0000 UTC m=+3698.597993022" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.192426 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.192511 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.192562 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.193925 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.193976 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" gracePeriod=600 Dec 05 10:08:20 crc kubenswrapper[4815]: E1205 10:08:20.449524 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:08:20 crc kubenswrapper[4815]: E1205 10:08:20.453169 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a874c44_831d_4f81_bf48_2025f5615574.slice/crio-9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1.scope\": RecentStats: unable to find data in memory cache]" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.644336 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.768379 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.768510 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-config-data\") pod \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.768547 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsfbs\" (UniqueName: \"kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-kube-api-access-nsfbs\") pod \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.768595 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-ceph\") pod \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.768661 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-combined-ca-bundle\") pod \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.768695 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-httpd-run\") pod \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.768833 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-logs\") pod \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.768873 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-internal-tls-certs\") pod \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.768917 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-scripts\") pod \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\" (UID: \"ee526317-f5a1-4682-b2ab-612ce15d9bf9\") " Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.773469 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-logs" (OuterVolumeSpecName: "logs") pod "ee526317-f5a1-4682-b2ab-612ce15d9bf9" (UID: "ee526317-f5a1-4682-b2ab-612ce15d9bf9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.773998 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ee526317-f5a1-4682-b2ab-612ce15d9bf9" (UID: "ee526317-f5a1-4682-b2ab-612ce15d9bf9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.793425 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" exitCode=0 Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.793504 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1"} Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.793538 4815 scope.go:117] "RemoveContainer" containerID="8af74f205ea1ac61f50c412f382831877290dbde40fda4bf5e34911bd0856599" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.794172 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:08:20 crc kubenswrapper[4815]: E1205 10:08:20.794519 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.799343 4815 generic.go:334] "Generic (PLEG): container finished" podID="ee526317-f5a1-4682-b2ab-612ce15d9bf9" containerID="3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39" exitCode=143 Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.799377 4815 generic.go:334] "Generic (PLEG): container finished" podID="ee526317-f5a1-4682-b2ab-612ce15d9bf9" containerID="a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510" exitCode=143 Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.799403 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee526317-f5a1-4682-b2ab-612ce15d9bf9","Type":"ContainerDied","Data":"3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39"} Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.799430 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee526317-f5a1-4682-b2ab-612ce15d9bf9","Type":"ContainerDied","Data":"a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510"} Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.799443 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"ee526317-f5a1-4682-b2ab-612ce15d9bf9","Type":"ContainerDied","Data":"125ea7302ad699d89f039c0b0631723a8f44d8efc69516d2ae3964113cef555c"} Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.799597 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.854789 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-ceph" (OuterVolumeSpecName: "ceph") pod "ee526317-f5a1-4682-b2ab-612ce15d9bf9" (UID: "ee526317-f5a1-4682-b2ab-612ce15d9bf9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.872039 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.872068 4815 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.872077 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ee526317-f5a1-4682-b2ab-612ce15d9bf9-logs\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.873288 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-kube-api-access-nsfbs" (OuterVolumeSpecName: "kube-api-access-nsfbs") pod "ee526317-f5a1-4682-b2ab-612ce15d9bf9" (UID: "ee526317-f5a1-4682-b2ab-612ce15d9bf9"). InnerVolumeSpecName "kube-api-access-nsfbs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.874711 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "ee526317-f5a1-4682-b2ab-612ce15d9bf9" (UID: "ee526317-f5a1-4682-b2ab-612ce15d9bf9"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.878137 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-scripts" (OuterVolumeSpecName: "scripts") pod "ee526317-f5a1-4682-b2ab-612ce15d9bf9" (UID: "ee526317-f5a1-4682-b2ab-612ce15d9bf9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.955923 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee526317-f5a1-4682-b2ab-612ce15d9bf9" (UID: "ee526317-f5a1-4682-b2ab-612ce15d9bf9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.973836 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.974133 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.974169 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.974179 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsfbs\" (UniqueName: \"kubernetes.io/projected/ee526317-f5a1-4682-b2ab-612ce15d9bf9-kube-api-access-nsfbs\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:20 crc kubenswrapper[4815]: I1205 10:08:20.985596 4815 scope.go:117] "RemoveContainer" containerID="3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.004970 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-config-data" (OuterVolumeSpecName: "config-data") pod "ee526317-f5a1-4682-b2ab-612ce15d9bf9" (UID: "ee526317-f5a1-4682-b2ab-612ce15d9bf9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.010971 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.038827 4815 scope.go:117] "RemoveContainer" containerID="a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.075793 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.075822 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.105701 4815 scope.go:117] "RemoveContainer" containerID="3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39" Dec 05 10:08:21 crc kubenswrapper[4815]: E1205 10:08:21.106260 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39\": container with ID starting with 3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39 not found: ID does not exist" containerID="3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.106299 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39"} err="failed to get container status \"3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39\": rpc error: code = NotFound desc = could not find container \"3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39\": container with ID starting with 3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39 not found: ID does not exist" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.106327 4815 scope.go:117] "RemoveContainer" containerID="a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510" Dec 05 10:08:21 crc kubenswrapper[4815]: E1205 10:08:21.106750 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510\": container with ID starting with a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510 not found: ID does not exist" containerID="a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.106778 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510"} err="failed to get container status \"a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510\": rpc error: code = NotFound desc = could not find container \"a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510\": container with ID starting with a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510 not found: ID does not exist" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.106795 4815 scope.go:117] "RemoveContainer" containerID="3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.106989 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39"} err="failed to get container status \"3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39\": rpc error: code = NotFound desc = could not find container \"3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39\": container with ID starting with 3a71740cab5e80f65136ccba354bcafa4372c11addceafeef3ec00b1c4148b39 not found: ID does not exist" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.107009 4815 scope.go:117] "RemoveContainer" containerID="a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.107202 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510"} err="failed to get container status \"a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510\": rpc error: code = NotFound desc = could not find container \"a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510\": container with ID starting with a567256f7553ed254cf1ef64e258086a2082e7092dc7e98b2b2ab6d3f9401510 not found: ID does not exist" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.123568 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.157397 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "ee526317-f5a1-4682-b2ab-612ce15d9bf9" (UID: "ee526317-f5a1-4682-b2ab-612ce15d9bf9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.177769 4815 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ee526317-f5a1-4682-b2ab-612ce15d9bf9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.524327 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.557691 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.606913 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:21 crc kubenswrapper[4815]: E1205 10:08:21.610178 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee526317-f5a1-4682-b2ab-612ce15d9bf9" containerName="glance-log" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.610197 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee526317-f5a1-4682-b2ab-612ce15d9bf9" containerName="glance-log" Dec 05 10:08:21 crc kubenswrapper[4815]: E1205 10:08:21.610212 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee526317-f5a1-4682-b2ab-612ce15d9bf9" containerName="glance-httpd" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.610221 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee526317-f5a1-4682-b2ab-612ce15d9bf9" containerName="glance-httpd" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.611232 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee526317-f5a1-4682-b2ab-612ce15d9bf9" containerName="glance-httpd" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.611264 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee526317-f5a1-4682-b2ab-612ce15d9bf9" containerName="glance-log" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.614071 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.617968 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.618180 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.663603 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.704716 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7330ea7b-3369-4ce7-95b6-fa415d3255f8-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.704773 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24hcz\" (UniqueName: \"kubernetes.io/projected/7330ea7b-3369-4ce7-95b6-fa415d3255f8-kube-api-access-24hcz\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.704798 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.704821 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7330ea7b-3369-4ce7-95b6-fa415d3255f8-logs\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.704849 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.704872 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.704928 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.704979 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7330ea7b-3369-4ce7-95b6-fa415d3255f8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.705030 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.806170 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.806267 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7330ea7b-3369-4ce7-95b6-fa415d3255f8-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.806298 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24hcz\" (UniqueName: \"kubernetes.io/projected/7330ea7b-3369-4ce7-95b6-fa415d3255f8-kube-api-access-24hcz\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.806326 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.806347 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7330ea7b-3369-4ce7-95b6-fa415d3255f8-logs\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.806372 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.806393 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.806459 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.806613 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7330ea7b-3369-4ce7-95b6-fa415d3255f8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.807031 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7330ea7b-3369-4ce7-95b6-fa415d3255f8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.808649 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.808833 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7330ea7b-3369-4ce7-95b6-fa415d3255f8-logs\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.813216 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7330ea7b-3369-4ce7-95b6-fa415d3255f8-ceph\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.813250 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.815228 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.817976 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.840128 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b","Type":"ContainerStarted","Data":"092da450fcfcb0af2a4be9f4bfacd15cd24db6d4f4352860886283e844a2237f"} Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.847368 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24hcz\" (UniqueName: \"kubernetes.io/projected/7330ea7b-3369-4ce7-95b6-fa415d3255f8-kube-api-access-24hcz\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.847672 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7330ea7b-3369-4ce7-95b6-fa415d3255f8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.857357 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-0\" (UID: \"7330ea7b-3369-4ce7-95b6-fa415d3255f8\") " pod="openstack/glance-default-internal-api-0" Dec 05 10:08:21 crc kubenswrapper[4815]: I1205 10:08:21.982251 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:22 crc kubenswrapper[4815]: I1205 10:08:22.180806 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 05 10:08:22 crc kubenswrapper[4815]: I1205 10:08:22.433106 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-volume-volume1-0" podUID="15d9387d-aef1-48fe-8fb1-cb9d3dec22ea" containerName="cinder-volume" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 05 10:08:22 crc kubenswrapper[4815]: I1205 10:08:22.495354 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 05 10:08:22 crc kubenswrapper[4815]: I1205 10:08:22.678897 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 05 10:08:22 crc kubenswrapper[4815]: W1205 10:08:22.716000 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7330ea7b_3369_4ce7_95b6_fa415d3255f8.slice/crio-8d6305fd7727c360d1dc4252fdbd786142db2adbafd112696803602808bd55d3 WatchSource:0}: Error finding container 8d6305fd7727c360d1dc4252fdbd786142db2adbafd112696803602808bd55d3: Status 404 returned error can't find the container with id 8d6305fd7727c360d1dc4252fdbd786142db2adbafd112696803602808bd55d3 Dec 05 10:08:22 crc kubenswrapper[4815]: I1205 10:08:22.927130 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b","Type":"ContainerStarted","Data":"e766b04c436e58b224f37f392acb155972e85f673bfc7c8e3e7701ae227a3d6e"} Dec 05 10:08:22 crc kubenswrapper[4815]: I1205 10:08:22.932871 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7330ea7b-3369-4ce7-95b6-fa415d3255f8","Type":"ContainerStarted","Data":"8d6305fd7727c360d1dc4252fdbd786142db2adbafd112696803602808bd55d3"} Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.302292 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-l5xf8"] Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.303770 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.307535 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-f2k2z" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.307593 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.346214 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-l5xf8"] Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.451385 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee526317-f5a1-4682-b2ab-612ce15d9bf9" path="/var/lib/kubelet/pods/ee526317-f5a1-4682-b2ab-612ce15d9bf9/volumes" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.478020 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-combined-ca-bundle\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.478078 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-job-config-data\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.478120 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-config-data\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.478319 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kbclp\" (UniqueName: \"kubernetes.io/projected/733c76db-715e-41b2-a91f-4032c25e2ab9-kube-api-access-kbclp\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.582738 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-combined-ca-bundle\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.582807 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-job-config-data\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.583702 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-config-data\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.584512 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kbclp\" (UniqueName: \"kubernetes.io/projected/733c76db-715e-41b2-a91f-4032c25e2ab9-kube-api-access-kbclp\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.602926 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-job-config-data\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.605300 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-combined-ca-bundle\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.605674 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-config-data\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.629053 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kbclp\" (UniqueName: \"kubernetes.io/projected/733c76db-715e-41b2-a91f-4032c25e2ab9-kube-api-access-kbclp\") pod \"manila-db-sync-l5xf8\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.652282 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:23 crc kubenswrapper[4815]: I1205 10:08:23.960280 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b","Type":"ContainerStarted","Data":"7cf23de71e15bdd51571b490104b481d48dc78a0a33fdffad524f573bee8b362"} Dec 05 10:08:24 crc kubenswrapper[4815]: I1205 10:08:24.690987 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.690968427 podStartE2EDuration="6.690968427s" podCreationTimestamp="2025-12-05 10:08:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:08:23.988899715 +0000 UTC m=+3702.867506552" watchObservedRunningTime="2025-12-05 10:08:24.690968427 +0000 UTC m=+3703.569575264" Dec 05 10:08:24 crc kubenswrapper[4815]: I1205 10:08:24.703973 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-l5xf8"] Dec 05 10:08:24 crc kubenswrapper[4815]: I1205 10:08:24.999395 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7330ea7b-3369-4ce7-95b6-fa415d3255f8","Type":"ContainerStarted","Data":"d4522bed653a494d23df9475f34041fdcde7d79a95eddaa377bd5560dbd71569"} Dec 05 10:08:26 crc kubenswrapper[4815]: I1205 10:08:26.038090 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7330ea7b-3369-4ce7-95b6-fa415d3255f8","Type":"ContainerStarted","Data":"f61e573647bae7c2b6be6e6fa5eeb48caf049e26c758e38eda9e6a95b3ab032d"} Dec 05 10:08:26 crc kubenswrapper[4815]: I1205 10:08:26.068335 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.068311184 podStartE2EDuration="5.068311184s" podCreationTimestamp="2025-12-05 10:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:08:26.056347318 +0000 UTC m=+3704.934954155" watchObservedRunningTime="2025-12-05 10:08:26.068311184 +0000 UTC m=+3704.946918021" Dec 05 10:08:27 crc kubenswrapper[4815]: I1205 10:08:27.063448 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 05 10:08:29 crc kubenswrapper[4815]: I1205 10:08:29.392787 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 10:08:29 crc kubenswrapper[4815]: I1205 10:08:29.393243 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 05 10:08:29 crc kubenswrapper[4815]: I1205 10:08:29.436146 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 10:08:29 crc kubenswrapper[4815]: I1205 10:08:29.436565 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 05 10:08:30 crc kubenswrapper[4815]: I1205 10:08:30.076590 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 10:08:30 crc kubenswrapper[4815]: I1205 10:08:30.076657 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 05 10:08:31 crc kubenswrapper[4815]: I1205 10:08:31.085836 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-l5xf8" event={"ID":"733c76db-715e-41b2-a91f-4032c25e2ab9","Type":"ContainerStarted","Data":"b0396425883f06eab04c72a78a24e3e5560e53aabc80cc40a60c562460f22e97"} Dec 05 10:08:31 crc kubenswrapper[4815]: I1205 10:08:31.982660 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:31 crc kubenswrapper[4815]: I1205 10:08:31.984189 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.073052 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.077116 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.111555 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cd8dc9dcc-bvxhw" event={"ID":"af69fb20-ecce-4a27-a4c1-5ff38416c9e1","Type":"ContainerStarted","Data":"b4e56f1e65152b015d2ec8e251f2d3499a6a3897df811494e2cc926704a8b74b"} Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.112355 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cd8dc9dcc-bvxhw" event={"ID":"af69fb20-ecce-4a27-a4c1-5ff38416c9e1","Type":"ContainerStarted","Data":"912b4cd5e30af3a7d368420e4626ce55e7db7da4339a186fde13ffa5d764b16c"} Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.112616 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6cd8dc9dcc-bvxhw" podUID="af69fb20-ecce-4a27-a4c1-5ff38416c9e1" containerName="horizon-log" containerID="cri-o://912b4cd5e30af3a7d368420e4626ce55e7db7da4339a186fde13ffa5d764b16c" gracePeriod=30 Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.112783 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6cd8dc9dcc-bvxhw" podUID="af69fb20-ecce-4a27-a4c1-5ff38416c9e1" containerName="horizon" containerID="cri-o://b4e56f1e65152b015d2ec8e251f2d3499a6a3897df811494e2cc926704a8b74b" gracePeriod=30 Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.122550 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cbc799c5c-zhk6n" event={"ID":"92a9f228-e917-4610-bc27-822d3e0e8578","Type":"ContainerStarted","Data":"e3fff7e3079a25bfa26dd2ef9ac3878bb5c60b83c22302b0d8e6f9ec5702eda9"} Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.122591 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cbc799c5c-zhk6n" event={"ID":"92a9f228-e917-4610-bc27-822d3e0e8578","Type":"ContainerStarted","Data":"6105960b1b3102457b5c2389017b5e626e34517696b07f97aac7e67b7c3474a1"} Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.122763 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-cbc799c5c-zhk6n" podUID="92a9f228-e917-4610-bc27-822d3e0e8578" containerName="horizon-log" containerID="cri-o://6105960b1b3102457b5c2389017b5e626e34517696b07f97aac7e67b7c3474a1" gracePeriod=30 Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.122826 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-cbc799c5c-zhk6n" podUID="92a9f228-e917-4610-bc27-822d3e0e8578" containerName="horizon" containerID="cri-o://e3fff7e3079a25bfa26dd2ef9ac3878bb5c60b83c22302b0d8e6f9ec5702eda9" gracePeriod=30 Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.128792 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c444bf898-b6z6t" event={"ID":"819b83db-6f8e-4b8d-81d7-719760d1ff6b","Type":"ContainerStarted","Data":"8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa"} Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.128830 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c444bf898-b6z6t" event={"ID":"819b83db-6f8e-4b8d-81d7-719760d1ff6b","Type":"ContainerStarted","Data":"d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b"} Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.137748 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74b8f449c4-9gwqr" event={"ID":"e9604337-767f-492a-aa51-f31578f44ece","Type":"ContainerStarted","Data":"51e679dcc1d48499552db92248835161fa848fe0a4cd210eae52627a536c8378"} Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.137829 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.137847 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-74b8f449c4-9gwqr" event={"ID":"e9604337-767f-492a-aa51-f31578f44ece","Type":"ContainerStarted","Data":"0b80294ef3c42a408878fb7582e0ecc412a3a9d0cf76fea1318e90ea62a81942"} Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.138102 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.138479 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6cd8dc9dcc-bvxhw" podStartSLOduration=2.926299532 podStartE2EDuration="20.138459548s" podCreationTimestamp="2025-12-05 10:08:12 +0000 UTC" firstStartedPulling="2025-12-05 10:08:14.046363786 +0000 UTC m=+3692.924970623" lastFinishedPulling="2025-12-05 10:08:31.258523802 +0000 UTC m=+3710.137130639" observedRunningTime="2025-12-05 10:08:32.133565754 +0000 UTC m=+3711.012172601" watchObservedRunningTime="2025-12-05 10:08:32.138459548 +0000 UTC m=+3711.017066385" Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.180204 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-cbc799c5c-zhk6n" podStartSLOduration=3.223320005 podStartE2EDuration="20.180187176s" podCreationTimestamp="2025-12-05 10:08:12 +0000 UTC" firstStartedPulling="2025-12-05 10:08:14.281640685 +0000 UTC m=+3693.160247522" lastFinishedPulling="2025-12-05 10:08:31.238507866 +0000 UTC m=+3710.117114693" observedRunningTime="2025-12-05 10:08:32.154854425 +0000 UTC m=+3711.033461262" watchObservedRunningTime="2025-12-05 10:08:32.180187176 +0000 UTC m=+3711.058794013" Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.188893 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-c444bf898-b6z6t" podStartSLOduration=3.470398531 podStartE2EDuration="17.188874003s" podCreationTimestamp="2025-12-05 10:08:15 +0000 UTC" firstStartedPulling="2025-12-05 10:08:17.566914592 +0000 UTC m=+3696.445521429" lastFinishedPulling="2025-12-05 10:08:31.285390064 +0000 UTC m=+3710.163996901" observedRunningTime="2025-12-05 10:08:32.178794028 +0000 UTC m=+3711.057400865" watchObservedRunningTime="2025-12-05 10:08:32.188874003 +0000 UTC m=+3711.067480840" Dec 05 10:08:32 crc kubenswrapper[4815]: I1205 10:08:32.209260 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-74b8f449c4-9gwqr" podStartSLOduration=2.795682385 podStartE2EDuration="16.209231898s" podCreationTimestamp="2025-12-05 10:08:16 +0000 UTC" firstStartedPulling="2025-12-05 10:08:17.89159394 +0000 UTC m=+3696.770200777" lastFinishedPulling="2025-12-05 10:08:31.305143443 +0000 UTC m=+3710.183750290" observedRunningTime="2025-12-05 10:08:32.198383902 +0000 UTC m=+3711.076990739" watchObservedRunningTime="2025-12-05 10:08:32.209231898 +0000 UTC m=+3711.087838735" Dec 05 10:08:33 crc kubenswrapper[4815]: I1205 10:08:33.301683 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:08:33 crc kubenswrapper[4815]: I1205 10:08:33.422408 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:08:33 crc kubenswrapper[4815]: E1205 10:08:33.422795 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:08:33 crc kubenswrapper[4815]: I1205 10:08:33.564873 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:08:36 crc kubenswrapper[4815]: I1205 10:08:36.484833 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:36 crc kubenswrapper[4815]: I1205 10:08:36.485322 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:08:36 crc kubenswrapper[4815]: I1205 10:08:36.902590 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:36 crc kubenswrapper[4815]: I1205 10:08:36.903839 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:08:38 crc kubenswrapper[4815]: I1205 10:08:38.307652 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:38 crc kubenswrapper[4815]: I1205 10:08:38.307773 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:08:38 crc kubenswrapper[4815]: I1205 10:08:38.330589 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 10:08:38 crc kubenswrapper[4815]: I1205 10:08:38.330732 4815 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 05 10:08:38 crc kubenswrapper[4815]: I1205 10:08:38.392196 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 05 10:08:38 crc kubenswrapper[4815]: I1205 10:08:38.538059 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 05 10:08:41 crc kubenswrapper[4815]: I1205 10:08:41.251000 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-l5xf8" event={"ID":"733c76db-715e-41b2-a91f-4032c25e2ab9","Type":"ContainerStarted","Data":"eebf0bf4b011d6ff3e9f4e7bcbe54282186f36f46fdebc15578a012747b13ee3"} Dec 05 10:08:41 crc kubenswrapper[4815]: I1205 10:08:41.287399 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-l5xf8" podStartSLOduration=8.589641984 podStartE2EDuration="18.287382004s" podCreationTimestamp="2025-12-05 10:08:23 +0000 UTC" firstStartedPulling="2025-12-05 10:08:30.909797017 +0000 UTC m=+3709.788403854" lastFinishedPulling="2025-12-05 10:08:40.607537037 +0000 UTC m=+3719.486143874" observedRunningTime="2025-12-05 10:08:41.284066294 +0000 UTC m=+3720.162673131" watchObservedRunningTime="2025-12-05 10:08:41.287382004 +0000 UTC m=+3720.165988841" Dec 05 10:08:46 crc kubenswrapper[4815]: I1205 10:08:46.486299 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c444bf898-b6z6t" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Dec 05 10:08:46 crc kubenswrapper[4815]: I1205 10:08:46.903409 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-74b8f449c4-9gwqr" podUID="e9604337-767f-492a-aa51-f31578f44ece" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.248:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.248:8443: connect: connection refused" Dec 05 10:08:48 crc kubenswrapper[4815]: I1205 10:08:48.418407 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:08:48 crc kubenswrapper[4815]: E1205 10:08:48.418716 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:08:54 crc kubenswrapper[4815]: I1205 10:08:54.363605 4815 generic.go:334] "Generic (PLEG): container finished" podID="733c76db-715e-41b2-a91f-4032c25e2ab9" containerID="eebf0bf4b011d6ff3e9f4e7bcbe54282186f36f46fdebc15578a012747b13ee3" exitCode=0 Dec 05 10:08:54 crc kubenswrapper[4815]: I1205 10:08:54.363692 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-l5xf8" event={"ID":"733c76db-715e-41b2-a91f-4032c25e2ab9","Type":"ContainerDied","Data":"eebf0bf4b011d6ff3e9f4e7bcbe54282186f36f46fdebc15578a012747b13ee3"} Dec 05 10:08:56 crc kubenswrapper[4815]: I1205 10:08:56.488392 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-c444bf898-b6z6t" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Dec 05 10:08:56 crc kubenswrapper[4815]: I1205 10:08:56.905356 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-74b8f449c4-9gwqr" podUID="e9604337-767f-492a-aa51-f31578f44ece" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.248:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.248:8443: connect: connection refused" Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.137336 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.296100 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kbclp\" (UniqueName: \"kubernetes.io/projected/733c76db-715e-41b2-a91f-4032c25e2ab9-kube-api-access-kbclp\") pod \"733c76db-715e-41b2-a91f-4032c25e2ab9\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.296197 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-job-config-data\") pod \"733c76db-715e-41b2-a91f-4032c25e2ab9\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.296280 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-config-data\") pod \"733c76db-715e-41b2-a91f-4032c25e2ab9\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.296591 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-combined-ca-bundle\") pod \"733c76db-715e-41b2-a91f-4032c25e2ab9\" (UID: \"733c76db-715e-41b2-a91f-4032c25e2ab9\") " Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.327333 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/733c76db-715e-41b2-a91f-4032c25e2ab9-kube-api-access-kbclp" (OuterVolumeSpecName: "kube-api-access-kbclp") pod "733c76db-715e-41b2-a91f-4032c25e2ab9" (UID: "733c76db-715e-41b2-a91f-4032c25e2ab9"). InnerVolumeSpecName "kube-api-access-kbclp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.328081 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "733c76db-715e-41b2-a91f-4032c25e2ab9" (UID: "733c76db-715e-41b2-a91f-4032c25e2ab9"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.329742 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-config-data" (OuterVolumeSpecName: "config-data") pod "733c76db-715e-41b2-a91f-4032c25e2ab9" (UID: "733c76db-715e-41b2-a91f-4032c25e2ab9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.367176 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "733c76db-715e-41b2-a91f-4032c25e2ab9" (UID: "733c76db-715e-41b2-a91f-4032c25e2ab9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.400177 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.400219 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kbclp\" (UniqueName: \"kubernetes.io/projected/733c76db-715e-41b2-a91f-4032c25e2ab9-kube-api-access-kbclp\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.400236 4815 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.400250 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/733c76db-715e-41b2-a91f-4032c25e2ab9-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.457598 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-l5xf8" event={"ID":"733c76db-715e-41b2-a91f-4032c25e2ab9","Type":"ContainerDied","Data":"b0396425883f06eab04c72a78a24e3e5560e53aabc80cc40a60c562460f22e97"} Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.457628 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-l5xf8" Dec 05 10:08:57 crc kubenswrapper[4815]: I1205 10:08:57.457636 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0396425883f06eab04c72a78a24e3e5560e53aabc80cc40a60c562460f22e97" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.444225 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 10:08:58 crc kubenswrapper[4815]: E1205 10:08:58.444922 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733c76db-715e-41b2-a91f-4032c25e2ab9" containerName="manila-db-sync" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.444947 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="733c76db-715e-41b2-a91f-4032c25e2ab9" containerName="manila-db-sync" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.445168 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="733c76db-715e-41b2-a91f-4032c25e2ab9" containerName="manila-db-sync" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.452209 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.458689 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.458712 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.458689 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-f2k2z" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.459075 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.480096 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.529922 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.530043 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9vss\" (UniqueName: \"kubernetes.io/projected/551a6e4e-9e8a-4aca-b569-966f6f562903-kube-api-access-s9vss\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.530070 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.530128 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/551a6e4e-9e8a-4aca-b569-966f6f562903-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.530143 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-scripts\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.530206 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.555213 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.556851 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.573814 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.585517 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.632271 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.632582 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.632708 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-scripts\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.632846 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb7bt\" (UniqueName: \"kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-kube-api-access-xb7bt\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.632967 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.633044 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.633136 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.633222 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9vss\" (UniqueName: \"kubernetes.io/projected/551a6e4e-9e8a-4aca-b569-966f6f562903-kube-api-access-s9vss\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.633302 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.633406 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-ceph\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.633503 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/551a6e4e-9e8a-4aca-b569-966f6f562903-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.634217 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-scripts\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.634360 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.634472 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.634237 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/551a6e4e-9e8a-4aca-b569-966f6f562903-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.640810 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.645184 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-scripts\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.646122 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.646693 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.680004 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9vss\" (UniqueName: \"kubernetes.io/projected/551a6e4e-9e8a-4aca-b569-966f6f562903-kube-api-access-s9vss\") pod \"manila-scheduler-0\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.735152 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.735452 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.735602 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-ceph\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.736482 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.736624 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.736775 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.736873 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-scripts\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.736974 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb7bt\" (UniqueName: \"kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-kube-api-access-xb7bt\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.737310 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.742271 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.743650 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.753419 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.755523 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-ceph\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.756088 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.760083 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-scripts\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.780144 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb7bt\" (UniqueName: \"kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-kube-api-access-xb7bt\") pod \"manila-share-share1-0\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.781437 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.815297 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-94c764695-ml6b2"] Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.817580 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.885795 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-94c764695-ml6b2"] Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.891324 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.941326 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-config\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.941368 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-ovsdbserver-nb\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.941436 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-ovsdbserver-sb\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.941537 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-openstack-edpm-ipam\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.941577 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl5gb\" (UniqueName: \"kubernetes.io/projected/310bad94-91d6-4420-9d0c-8ed3af8b72c5-kube-api-access-tl5gb\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.942869 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-dns-svc\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.972573 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.974314 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.980686 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 05 10:08:58 crc kubenswrapper[4815]: I1205 10:08:58.988520 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.044562 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-dns-svc\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.044640 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-config\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.044659 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-ovsdbserver-nb\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.044715 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-ovsdbserver-sb\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.044750 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-openstack-edpm-ipam\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.044805 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl5gb\" (UniqueName: \"kubernetes.io/projected/310bad94-91d6-4420-9d0c-8ed3af8b72c5-kube-api-access-tl5gb\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.045874 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-config\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.046462 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-ovsdbserver-nb\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.047391 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-openstack-edpm-ipam\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.051200 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-dns-svc\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.053540 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/310bad94-91d6-4420-9d0c-8ed3af8b72c5-ovsdbserver-sb\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.078927 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl5gb\" (UniqueName: \"kubernetes.io/projected/310bad94-91d6-4420-9d0c-8ed3af8b72c5-kube-api-access-tl5gb\") pod \"dnsmasq-dns-94c764695-ml6b2\" (UID: \"310bad94-91d6-4420-9d0c-8ed3af8b72c5\") " pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.146481 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-scripts\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.146640 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64241d4c-dcb2-4c6b-9963-f2c61f586826-etc-machine-id\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.147803 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data-custom\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.147914 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvp9s\" (UniqueName: \"kubernetes.io/projected/64241d4c-dcb2-4c6b-9963-f2c61f586826-kube-api-access-qvp9s\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.148018 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.148091 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.148174 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64241d4c-dcb2-4c6b-9963-f2c61f586826-logs\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.159939 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.250678 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64241d4c-dcb2-4c6b-9963-f2c61f586826-etc-machine-id\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.250723 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data-custom\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.250758 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvp9s\" (UniqueName: \"kubernetes.io/projected/64241d4c-dcb2-4c6b-9963-f2c61f586826-kube-api-access-qvp9s\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.250776 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.250791 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.250822 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64241d4c-dcb2-4c6b-9963-f2c61f586826-logs\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.250824 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64241d4c-dcb2-4c6b-9963-f2c61f586826-etc-machine-id\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.250979 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-scripts\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.251254 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64241d4c-dcb2-4c6b-9963-f2c61f586826-logs\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.254431 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data-custom\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.256400 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-scripts\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.258228 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.399481 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.416224 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvp9s\" (UniqueName: \"kubernetes.io/projected/64241d4c-dcb2-4c6b-9963-f2c61f586826-kube-api-access-qvp9s\") pod \"manila-api-0\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " pod="openstack/manila-api-0" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.431660 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:08:59 crc kubenswrapper[4815]: E1205 10:08:59.433020 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:08:59 crc kubenswrapper[4815]: I1205 10:08:59.614532 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 05 10:09:00 crc kubenswrapper[4815]: I1205 10:09:00.172487 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 10:09:00 crc kubenswrapper[4815]: I1205 10:09:00.229958 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-94c764695-ml6b2"] Dec 05 10:09:00 crc kubenswrapper[4815]: I1205 10:09:00.241552 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 10:09:00 crc kubenswrapper[4815]: I1205 10:09:00.599363 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 05 10:09:00 crc kubenswrapper[4815]: I1205 10:09:00.601995 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"944629dc-36f8-4784-ac9c-7dea31445d5a","Type":"ContainerStarted","Data":"d234426f46b59747c9df7c1678726b4079d5c7e2ad7d94d913cf5eaeb22c20d2"} Dec 05 10:09:00 crc kubenswrapper[4815]: I1205 10:09:00.603907 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-94c764695-ml6b2" event={"ID":"310bad94-91d6-4420-9d0c-8ed3af8b72c5","Type":"ContainerStarted","Data":"5a66df65b5a9c54838a8317cad7dee549b3eeb661e6576b4fb741f4833dda086"} Dec 05 10:09:00 crc kubenswrapper[4815]: I1205 10:09:00.603934 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-94c764695-ml6b2" event={"ID":"310bad94-91d6-4420-9d0c-8ed3af8b72c5","Type":"ContainerStarted","Data":"196120256e1305a9a6c3f4b7a573a57d697d82cad6f167feda48b85bf502fef5"} Dec 05 10:09:00 crc kubenswrapper[4815]: I1205 10:09:00.605300 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"551a6e4e-9e8a-4aca-b569-966f6f562903","Type":"ContainerStarted","Data":"97a41535c7b9f258138807d847e89332e3b78b2da7139318526d3417acef3e45"} Dec 05 10:09:01 crc kubenswrapper[4815]: I1205 10:09:01.633679 4815 generic.go:334] "Generic (PLEG): container finished" podID="310bad94-91d6-4420-9d0c-8ed3af8b72c5" containerID="5a66df65b5a9c54838a8317cad7dee549b3eeb661e6576b4fb741f4833dda086" exitCode=0 Dec 05 10:09:01 crc kubenswrapper[4815]: I1205 10:09:01.634172 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-94c764695-ml6b2" event={"ID":"310bad94-91d6-4420-9d0c-8ed3af8b72c5","Type":"ContainerDied","Data":"5a66df65b5a9c54838a8317cad7dee549b3eeb661e6576b4fb741f4833dda086"} Dec 05 10:09:01 crc kubenswrapper[4815]: I1205 10:09:01.634200 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-94c764695-ml6b2" event={"ID":"310bad94-91d6-4420-9d0c-8ed3af8b72c5","Type":"ContainerStarted","Data":"354e51310c78298453b04bfb2c9d2b77001ce91775002273144321d774b05940"} Dec 05 10:09:01 crc kubenswrapper[4815]: I1205 10:09:01.634250 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:09:01 crc kubenswrapper[4815]: I1205 10:09:01.646379 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"64241d4c-dcb2-4c6b-9963-f2c61f586826","Type":"ContainerStarted","Data":"9b5ef59b7c76a8e7ff21ff845139f758ff4fb02a06bb818f2220d826cc1f0ee5"} Dec 05 10:09:01 crc kubenswrapper[4815]: I1205 10:09:01.646435 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"64241d4c-dcb2-4c6b-9963-f2c61f586826","Type":"ContainerStarted","Data":"c4cd7d754dd647e8d94d207028f851d75f9b11716bfdedd1ad49f832925d852c"} Dec 05 10:09:01 crc kubenswrapper[4815]: I1205 10:09:01.665766 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-94c764695-ml6b2" podStartSLOduration=3.665737248 podStartE2EDuration="3.665737248s" podCreationTimestamp="2025-12-05 10:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:09:01.658381348 +0000 UTC m=+3740.536988185" watchObservedRunningTime="2025-12-05 10:09:01.665737248 +0000 UTC m=+3740.544344085" Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.306116 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.725817 4815 generic.go:334] "Generic (PLEG): container finished" podID="af69fb20-ecce-4a27-a4c1-5ff38416c9e1" containerID="b4e56f1e65152b015d2ec8e251f2d3499a6a3897df811494e2cc926704a8b74b" exitCode=137 Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.726374 4815 generic.go:334] "Generic (PLEG): container finished" podID="af69fb20-ecce-4a27-a4c1-5ff38416c9e1" containerID="912b4cd5e30af3a7d368420e4626ce55e7db7da4339a186fde13ffa5d764b16c" exitCode=137 Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.726458 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cd8dc9dcc-bvxhw" event={"ID":"af69fb20-ecce-4a27-a4c1-5ff38416c9e1","Type":"ContainerDied","Data":"b4e56f1e65152b015d2ec8e251f2d3499a6a3897df811494e2cc926704a8b74b"} Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.735582 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cd8dc9dcc-bvxhw" event={"ID":"af69fb20-ecce-4a27-a4c1-5ff38416c9e1","Type":"ContainerDied","Data":"912b4cd5e30af3a7d368420e4626ce55e7db7da4339a186fde13ffa5d764b16c"} Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.737690 4815 generic.go:334] "Generic (PLEG): container finished" podID="92a9f228-e917-4610-bc27-822d3e0e8578" containerID="e3fff7e3079a25bfa26dd2ef9ac3878bb5c60b83c22302b0d8e6f9ec5702eda9" exitCode=137 Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.737727 4815 generic.go:334] "Generic (PLEG): container finished" podID="92a9f228-e917-4610-bc27-822d3e0e8578" containerID="6105960b1b3102457b5c2389017b5e626e34517696b07f97aac7e67b7c3474a1" exitCode=137 Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.737786 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cbc799c5c-zhk6n" event={"ID":"92a9f228-e917-4610-bc27-822d3e0e8578","Type":"ContainerDied","Data":"e3fff7e3079a25bfa26dd2ef9ac3878bb5c60b83c22302b0d8e6f9ec5702eda9"} Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.737814 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cbc799c5c-zhk6n" event={"ID":"92a9f228-e917-4610-bc27-822d3e0e8578","Type":"ContainerDied","Data":"6105960b1b3102457b5c2389017b5e626e34517696b07f97aac7e67b7c3474a1"} Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.742080 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"551a6e4e-9e8a-4aca-b569-966f6f562903","Type":"ContainerStarted","Data":"5f00bf238719fb0032f9be6155c09e2a5b585fcf3cdedd6e29b08ed3e0fb3556"} Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.748912 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="64241d4c-dcb2-4c6b-9963-f2c61f586826" containerName="manila-api-log" containerID="cri-o://9b5ef59b7c76a8e7ff21ff845139f758ff4fb02a06bb818f2220d826cc1f0ee5" gracePeriod=30 Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.749764 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"64241d4c-dcb2-4c6b-9963-f2c61f586826","Type":"ContainerStarted","Data":"cd4f60812087b29766771717e3ac3332d8a1b3f5c4ffff1f7f2a9bca614c12e0"} Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.749857 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="64241d4c-dcb2-4c6b-9963-f2c61f586826" containerName="manila-api" containerID="cri-o://cd4f60812087b29766771717e3ac3332d8a1b3f5c4ffff1f7f2a9bca614c12e0" gracePeriod=30 Dec 05 10:09:02 crc kubenswrapper[4815]: I1205 10:09:02.751426 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.011219 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=5.011199534 podStartE2EDuration="5.011199534s" podCreationTimestamp="2025-12-05 10:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:09:03.00918961 +0000 UTC m=+3741.887796447" watchObservedRunningTime="2025-12-05 10:09:03.011199534 +0000 UTC m=+3741.889806371" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.225715 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:09:03 crc kubenswrapper[4815]: E1205 10:09:03.352991 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92a9f228_e917_4610_bc27_822d3e0e8578.slice/crio-conmon-e3fff7e3079a25bfa26dd2ef9ac3878bb5c60b83c22302b0d8e6f9ec5702eda9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf69fb20_ecce_4a27_a4c1_5ff38416c9e1.slice/crio-conmon-b4e56f1e65152b015d2ec8e251f2d3499a6a3897df811494e2cc926704a8b74b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf69fb20_ecce_4a27_a4c1_5ff38416c9e1.slice/crio-b4e56f1e65152b015d2ec8e251f2d3499a6a3897df811494e2cc926704a8b74b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64241d4c_dcb2_4c6b_9963_f2c61f586826.slice/crio-conmon-9b5ef59b7c76a8e7ff21ff845139f758ff4fb02a06bb818f2220d826cc1f0ee5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64241d4c_dcb2_4c6b_9963_f2c61f586826.slice/crio-cd4f60812087b29766771717e3ac3332d8a1b3f5c4ffff1f7f2a9bca614c12e0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod64241d4c_dcb2_4c6b_9963_f2c61f586826.slice/crio-9b5ef59b7c76a8e7ff21ff845139f758ff4fb02a06bb818f2220d826cc1f0ee5.scope\": RecentStats: unable to find data in memory cache]" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.400236 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-config-data\") pod \"92a9f228-e917-4610-bc27-822d3e0e8578\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.400295 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsqhf\" (UniqueName: \"kubernetes.io/projected/92a9f228-e917-4610-bc27-822d3e0e8578-kube-api-access-xsqhf\") pod \"92a9f228-e917-4610-bc27-822d3e0e8578\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.400367 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-scripts\") pod \"92a9f228-e917-4610-bc27-822d3e0e8578\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.400421 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/92a9f228-e917-4610-bc27-822d3e0e8578-horizon-secret-key\") pod \"92a9f228-e917-4610-bc27-822d3e0e8578\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.400442 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92a9f228-e917-4610-bc27-822d3e0e8578-logs\") pod \"92a9f228-e917-4610-bc27-822d3e0e8578\" (UID: \"92a9f228-e917-4610-bc27-822d3e0e8578\") " Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.406139 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92a9f228-e917-4610-bc27-822d3e0e8578-logs" (OuterVolumeSpecName: "logs") pod "92a9f228-e917-4610-bc27-822d3e0e8578" (UID: "92a9f228-e917-4610-bc27-822d3e0e8578"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.411185 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92a9f228-e917-4610-bc27-822d3e0e8578-kube-api-access-xsqhf" (OuterVolumeSpecName: "kube-api-access-xsqhf") pod "92a9f228-e917-4610-bc27-822d3e0e8578" (UID: "92a9f228-e917-4610-bc27-822d3e0e8578"). InnerVolumeSpecName "kube-api-access-xsqhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.425623 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92a9f228-e917-4610-bc27-822d3e0e8578-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "92a9f228-e917-4610-bc27-822d3e0e8578" (UID: "92a9f228-e917-4610-bc27-822d3e0e8578"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.447236 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-scripts" (OuterVolumeSpecName: "scripts") pod "92a9f228-e917-4610-bc27-822d3e0e8578" (UID: "92a9f228-e917-4610-bc27-822d3e0e8578"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.481748 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-config-data" (OuterVolumeSpecName: "config-data") pod "92a9f228-e917-4610-bc27-822d3e0e8578" (UID: "92a9f228-e917-4610-bc27-822d3e0e8578"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.482391 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.503930 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.504064 4815 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/92a9f228-e917-4610-bc27-822d3e0e8578-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.504140 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/92a9f228-e917-4610-bc27-822d3e0e8578-logs\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.504198 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/92a9f228-e917-4610-bc27-822d3e0e8578-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.504250 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsqhf\" (UniqueName: \"kubernetes.io/projected/92a9f228-e917-4610-bc27-822d3e0e8578-kube-api-access-xsqhf\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.606019 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-horizon-secret-key\") pod \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.606127 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-scripts\") pod \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.606232 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-logs\") pod \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.606285 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sc8n5\" (UniqueName: \"kubernetes.io/projected/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-kube-api-access-sc8n5\") pod \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.606330 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-config-data\") pod \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\" (UID: \"af69fb20-ecce-4a27-a4c1-5ff38416c9e1\") " Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.612898 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-logs" (OuterVolumeSpecName: "logs") pod "af69fb20-ecce-4a27-a4c1-5ff38416c9e1" (UID: "af69fb20-ecce-4a27-a4c1-5ff38416c9e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.616388 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-kube-api-access-sc8n5" (OuterVolumeSpecName: "kube-api-access-sc8n5") pod "af69fb20-ecce-4a27-a4c1-5ff38416c9e1" (UID: "af69fb20-ecce-4a27-a4c1-5ff38416c9e1"). InnerVolumeSpecName "kube-api-access-sc8n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.617610 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "af69fb20-ecce-4a27-a4c1-5ff38416c9e1" (UID: "af69fb20-ecce-4a27-a4c1-5ff38416c9e1"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.636230 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-scripts" (OuterVolumeSpecName: "scripts") pod "af69fb20-ecce-4a27-a4c1-5ff38416c9e1" (UID: "af69fb20-ecce-4a27-a4c1-5ff38416c9e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.659276 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-config-data" (OuterVolumeSpecName: "config-data") pod "af69fb20-ecce-4a27-a4c1-5ff38416c9e1" (UID: "af69fb20-ecce-4a27-a4c1-5ff38416c9e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.709665 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-logs\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.709711 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sc8n5\" (UniqueName: \"kubernetes.io/projected/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-kube-api-access-sc8n5\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.709723 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.709731 4815 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.709740 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/af69fb20-ecce-4a27-a4c1-5ff38416c9e1-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.801807 4815 generic.go:334] "Generic (PLEG): container finished" podID="64241d4c-dcb2-4c6b-9963-f2c61f586826" containerID="cd4f60812087b29766771717e3ac3332d8a1b3f5c4ffff1f7f2a9bca614c12e0" exitCode=143 Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.801840 4815 generic.go:334] "Generic (PLEG): container finished" podID="64241d4c-dcb2-4c6b-9963-f2c61f586826" containerID="9b5ef59b7c76a8e7ff21ff845139f758ff4fb02a06bb818f2220d826cc1f0ee5" exitCode=143 Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.801894 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"64241d4c-dcb2-4c6b-9963-f2c61f586826","Type":"ContainerDied","Data":"cd4f60812087b29766771717e3ac3332d8a1b3f5c4ffff1f7f2a9bca614c12e0"} Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.801921 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"64241d4c-dcb2-4c6b-9963-f2c61f586826","Type":"ContainerDied","Data":"9b5ef59b7c76a8e7ff21ff845139f758ff4fb02a06bb818f2220d826cc1f0ee5"} Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.803555 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6cd8dc9dcc-bvxhw" event={"ID":"af69fb20-ecce-4a27-a4c1-5ff38416c9e1","Type":"ContainerDied","Data":"5c63fdc252f9135984102cac80fddb6c5684fab2c7fb550cd2a08de7282e78d9"} Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.803586 4815 scope.go:117] "RemoveContainer" containerID="b4e56f1e65152b015d2ec8e251f2d3499a6a3897df811494e2cc926704a8b74b" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.803702 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6cd8dc9dcc-bvxhw" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.823023 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-cbc799c5c-zhk6n" event={"ID":"92a9f228-e917-4610-bc27-822d3e0e8578","Type":"ContainerDied","Data":"9fd67bce401ba820ad1cae4a96dfe211cba0e5902cc5fd1546f6691205f9c671"} Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.824121 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-cbc799c5c-zhk6n" Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.845542 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"551a6e4e-9e8a-4aca-b569-966f6f562903","Type":"ContainerStarted","Data":"ba8ba55cd835b36ab2fa6a0c5df436127112db8452432ecc3007ad8a7ceb9e2b"} Dec 05 10:09:03 crc kubenswrapper[4815]: I1205 10:09:03.896853 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=5.094834397 podStartE2EDuration="5.896830177s" podCreationTimestamp="2025-12-05 10:08:58 +0000 UTC" firstStartedPulling="2025-12-05 10:09:00.246656374 +0000 UTC m=+3739.125263211" lastFinishedPulling="2025-12-05 10:09:01.048652164 +0000 UTC m=+3739.927258991" observedRunningTime="2025-12-05 10:09:03.871192956 +0000 UTC m=+3742.749799793" watchObservedRunningTime="2025-12-05 10:09:03.896830177 +0000 UTC m=+3742.775437014" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.016611 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6cd8dc9dcc-bvxhw"] Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.022113 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.029445 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6cd8dc9dcc-bvxhw"] Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.039211 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-cbc799c5c-zhk6n"] Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.047293 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-cbc799c5c-zhk6n"] Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.129382 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data-custom\") pod \"64241d4c-dcb2-4c6b-9963-f2c61f586826\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.129679 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data\") pod \"64241d4c-dcb2-4c6b-9963-f2c61f586826\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.129769 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvp9s\" (UniqueName: \"kubernetes.io/projected/64241d4c-dcb2-4c6b-9963-f2c61f586826-kube-api-access-qvp9s\") pod \"64241d4c-dcb2-4c6b-9963-f2c61f586826\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.129875 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-combined-ca-bundle\") pod \"64241d4c-dcb2-4c6b-9963-f2c61f586826\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.130009 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-scripts\") pod \"64241d4c-dcb2-4c6b-9963-f2c61f586826\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.130082 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64241d4c-dcb2-4c6b-9963-f2c61f586826-logs\") pod \"64241d4c-dcb2-4c6b-9963-f2c61f586826\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.130177 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64241d4c-dcb2-4c6b-9963-f2c61f586826-etc-machine-id\") pod \"64241d4c-dcb2-4c6b-9963-f2c61f586826\" (UID: \"64241d4c-dcb2-4c6b-9963-f2c61f586826\") " Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.132911 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/64241d4c-dcb2-4c6b-9963-f2c61f586826-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "64241d4c-dcb2-4c6b-9963-f2c61f586826" (UID: "64241d4c-dcb2-4c6b-9963-f2c61f586826"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.132937 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64241d4c-dcb2-4c6b-9963-f2c61f586826-logs" (OuterVolumeSpecName: "logs") pod "64241d4c-dcb2-4c6b-9963-f2c61f586826" (UID: "64241d4c-dcb2-4c6b-9963-f2c61f586826"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.141626 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "64241d4c-dcb2-4c6b-9963-f2c61f586826" (UID: "64241d4c-dcb2-4c6b-9963-f2c61f586826"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.148870 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64241d4c-dcb2-4c6b-9963-f2c61f586826-kube-api-access-qvp9s" (OuterVolumeSpecName: "kube-api-access-qvp9s") pod "64241d4c-dcb2-4c6b-9963-f2c61f586826" (UID: "64241d4c-dcb2-4c6b-9963-f2c61f586826"). InnerVolumeSpecName "kube-api-access-qvp9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.158638 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-scripts" (OuterVolumeSpecName: "scripts") pod "64241d4c-dcb2-4c6b-9963-f2c61f586826" (UID: "64241d4c-dcb2-4c6b-9963-f2c61f586826"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.181618 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64241d4c-dcb2-4c6b-9963-f2c61f586826" (UID: "64241d4c-dcb2-4c6b-9963-f2c61f586826"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.240709 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvp9s\" (UniqueName: \"kubernetes.io/projected/64241d4c-dcb2-4c6b-9963-f2c61f586826-kube-api-access-qvp9s\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.240734 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.240746 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.240754 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64241d4c-dcb2-4c6b-9963-f2c61f586826-logs\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.240763 4815 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/64241d4c-dcb2-4c6b-9963-f2c61f586826-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.240773 4815 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.259297 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data" (OuterVolumeSpecName: "config-data") pod "64241d4c-dcb2-4c6b-9963-f2c61f586826" (UID: "64241d4c-dcb2-4c6b-9963-f2c61f586826"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.311562 4815 scope.go:117] "RemoveContainer" containerID="912b4cd5e30af3a7d368420e4626ce55e7db7da4339a186fde13ffa5d764b16c" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.500801 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64241d4c-dcb2-4c6b-9963-f2c61f586826-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.504383 4815 scope.go:117] "RemoveContainer" containerID="e3fff7e3079a25bfa26dd2ef9ac3878bb5c60b83c22302b0d8e6f9ec5702eda9" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.701814 4815 scope.go:117] "RemoveContainer" containerID="6105960b1b3102457b5c2389017b5e626e34517696b07f97aac7e67b7c3474a1" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.866945 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"64241d4c-dcb2-4c6b-9963-f2c61f586826","Type":"ContainerDied","Data":"c4cd7d754dd647e8d94d207028f851d75f9b11716bfdedd1ad49f832925d852c"} Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.866997 4815 scope.go:117] "RemoveContainer" containerID="cd4f60812087b29766771717e3ac3332d8a1b3f5c4ffff1f7f2a9bca614c12e0" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.866968 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.927674 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.940623 4815 scope.go:117] "RemoveContainer" containerID="9b5ef59b7c76a8e7ff21ff845139f758ff4fb02a06bb818f2220d826cc1f0ee5" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.958180 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.979582 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 05 10:09:04 crc kubenswrapper[4815]: E1205 10:09:04.980118 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a9f228-e917-4610-bc27-822d3e0e8578" containerName="horizon-log" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980137 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a9f228-e917-4610-bc27-822d3e0e8578" containerName="horizon-log" Dec 05 10:09:04 crc kubenswrapper[4815]: E1205 10:09:04.980151 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af69fb20-ecce-4a27-a4c1-5ff38416c9e1" containerName="horizon" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980157 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="af69fb20-ecce-4a27-a4c1-5ff38416c9e1" containerName="horizon" Dec 05 10:09:04 crc kubenswrapper[4815]: E1205 10:09:04.980168 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64241d4c-dcb2-4c6b-9963-f2c61f586826" containerName="manila-api-log" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980174 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="64241d4c-dcb2-4c6b-9963-f2c61f586826" containerName="manila-api-log" Dec 05 10:09:04 crc kubenswrapper[4815]: E1205 10:09:04.980183 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64241d4c-dcb2-4c6b-9963-f2c61f586826" containerName="manila-api" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980189 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="64241d4c-dcb2-4c6b-9963-f2c61f586826" containerName="manila-api" Dec 05 10:09:04 crc kubenswrapper[4815]: E1205 10:09:04.980206 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92a9f228-e917-4610-bc27-822d3e0e8578" containerName="horizon" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980211 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="92a9f228-e917-4610-bc27-822d3e0e8578" containerName="horizon" Dec 05 10:09:04 crc kubenswrapper[4815]: E1205 10:09:04.980224 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af69fb20-ecce-4a27-a4c1-5ff38416c9e1" containerName="horizon-log" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980229 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="af69fb20-ecce-4a27-a4c1-5ff38416c9e1" containerName="horizon-log" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980389 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="af69fb20-ecce-4a27-a4c1-5ff38416c9e1" containerName="horizon" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980405 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="64241d4c-dcb2-4c6b-9963-f2c61f586826" containerName="manila-api" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980418 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="64241d4c-dcb2-4c6b-9963-f2c61f586826" containerName="manila-api-log" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980426 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a9f228-e917-4610-bc27-822d3e0e8578" containerName="horizon-log" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980439 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="af69fb20-ecce-4a27-a4c1-5ff38416c9e1" containerName="horizon-log" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.980450 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="92a9f228-e917-4610-bc27-822d3e0e8578" containerName="horizon" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.981454 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.987750 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.988101 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.988158 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Dec 05 10:09:04 crc kubenswrapper[4815]: I1205 10:09:04.989368 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.010867 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-internal-tls-certs\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.010953 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/de4b5d94-2c11-4552-8230-08b776caa2a8-etc-machine-id\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.010973 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-config-data\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.010994 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-config-data-custom\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.011015 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68b6j\" (UniqueName: \"kubernetes.io/projected/de4b5d94-2c11-4552-8230-08b776caa2a8-kube-api-access-68b6j\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.011038 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-public-tls-certs\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.011066 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-scripts\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.011117 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.011147 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de4b5d94-2c11-4552-8230-08b776caa2a8-logs\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.117255 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68b6j\" (UniqueName: \"kubernetes.io/projected/de4b5d94-2c11-4552-8230-08b776caa2a8-kube-api-access-68b6j\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.117632 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-public-tls-certs\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.117791 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-scripts\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.117949 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.118064 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de4b5d94-2c11-4552-8230-08b776caa2a8-logs\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.118243 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-internal-tls-certs\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.118395 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/de4b5d94-2c11-4552-8230-08b776caa2a8-etc-machine-id\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.118510 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-config-data\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.118607 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-config-data-custom\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.121691 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/de4b5d94-2c11-4552-8230-08b776caa2a8-etc-machine-id\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.122014 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de4b5d94-2c11-4552-8230-08b776caa2a8-logs\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.127589 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-public-tls-certs\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.130125 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-config-data\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.131034 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-config-data-custom\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.131201 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-internal-tls-certs\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.139061 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-scripts\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.144518 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de4b5d94-2c11-4552-8230-08b776caa2a8-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.168113 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68b6j\" (UniqueName: \"kubernetes.io/projected/de4b5d94-2c11-4552-8230-08b776caa2a8-kube-api-access-68b6j\") pod \"manila-api-0\" (UID: \"de4b5d94-2c11-4552-8230-08b776caa2a8\") " pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.327278 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.486683 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64241d4c-dcb2-4c6b-9963-f2c61f586826" path="/var/lib/kubelet/pods/64241d4c-dcb2-4c6b-9963-f2c61f586826/volumes" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.491162 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92a9f228-e917-4610-bc27-822d3e0e8578" path="/var/lib/kubelet/pods/92a9f228-e917-4610-bc27-822d3e0e8578/volumes" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.492502 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af69fb20-ecce-4a27-a4c1-5ff38416c9e1" path="/var/lib/kubelet/pods/af69fb20-ecce-4a27-a4c1-5ff38416c9e1/volumes" Dec 05 10:09:05 crc kubenswrapper[4815]: I1205 10:09:05.980417 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 05 10:09:05 crc kubenswrapper[4815]: W1205 10:09:05.987289 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podde4b5d94_2c11_4552_8230_08b776caa2a8.slice/crio-97ad9972bc8c098c0965d42120fc4aa7857871bec49dd4f14293e80ba37bea72 WatchSource:0}: Error finding container 97ad9972bc8c098c0965d42120fc4aa7857871bec49dd4f14293e80ba37bea72: Status 404 returned error can't find the container with id 97ad9972bc8c098c0965d42120fc4aa7857871bec49dd4f14293e80ba37bea72 Dec 05 10:09:06 crc kubenswrapper[4815]: I1205 10:09:06.920083 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"de4b5d94-2c11-4552-8230-08b776caa2a8","Type":"ContainerStarted","Data":"97ad9972bc8c098c0965d42120fc4aa7857871bec49dd4f14293e80ba37bea72"} Dec 05 10:09:07 crc kubenswrapper[4815]: I1205 10:09:07.938267 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"de4b5d94-2c11-4552-8230-08b776caa2a8","Type":"ContainerStarted","Data":"0dea590b28cab3d408c4ce9581e82ab34329e7096fdc9e4de5eb3bbff2df3219"} Dec 05 10:09:07 crc kubenswrapper[4815]: I1205 10:09:07.938850 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 05 10:09:07 crc kubenswrapper[4815]: I1205 10:09:07.938885 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"de4b5d94-2c11-4552-8230-08b776caa2a8","Type":"ContainerStarted","Data":"0d591c057d739ab0b439be160595821a9511087af5f2d08ce1c9bf3b7802bbb6"} Dec 05 10:09:07 crc kubenswrapper[4815]: I1205 10:09:07.969055 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.969030782 podStartE2EDuration="3.969030782s" podCreationTimestamp="2025-12-05 10:09:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:09:07.959459562 +0000 UTC m=+3746.838066409" watchObservedRunningTime="2025-12-05 10:09:07.969030782 +0000 UTC m=+3746.847637629" Dec 05 10:09:08 crc kubenswrapper[4815]: I1205 10:09:08.782830 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 05 10:09:09 crc kubenswrapper[4815]: I1205 10:09:09.151794 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:09:09 crc kubenswrapper[4815]: I1205 10:09:09.165851 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-94c764695-ml6b2" Dec 05 10:09:09 crc kubenswrapper[4815]: I1205 10:09:09.272229 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54fcc757dc-kbs2f"] Dec 05 10:09:09 crc kubenswrapper[4815]: I1205 10:09:09.277633 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" podUID="a06a9288-ec09-4ff0-8a21-c0a131880da8" containerName="dnsmasq-dns" containerID="cri-o://ee9878b701d75509d14856c88dd79be3eef3919f11c76c7f390446563bd86c7e" gracePeriod=10 Dec 05 10:09:09 crc kubenswrapper[4815]: I1205 10:09:09.509104 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:09:09 crc kubenswrapper[4815]: I1205 10:09:09.960776 4815 generic.go:334] "Generic (PLEG): container finished" podID="a06a9288-ec09-4ff0-8a21-c0a131880da8" containerID="ee9878b701d75509d14856c88dd79be3eef3919f11c76c7f390446563bd86c7e" exitCode=0 Dec 05 10:09:09 crc kubenswrapper[4815]: I1205 10:09:09.960969 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" event={"ID":"a06a9288-ec09-4ff0-8a21-c0a131880da8","Type":"ContainerDied","Data":"ee9878b701d75509d14856c88dd79be3eef3919f11c76c7f390446563bd86c7e"} Dec 05 10:09:10 crc kubenswrapper[4815]: I1205 10:09:10.182478 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" podUID="a06a9288-ec09-4ff0-8a21-c0a131880da8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.194:5353: connect: connection refused" Dec 05 10:09:10 crc kubenswrapper[4815]: I1205 10:09:10.419283 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:09:10 crc kubenswrapper[4815]: E1205 10:09:10.419597 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:09:11 crc kubenswrapper[4815]: I1205 10:09:11.696278 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-74b8f449c4-9gwqr" Dec 05 10:09:11 crc kubenswrapper[4815]: I1205 10:09:11.702198 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:09:11 crc kubenswrapper[4815]: I1205 10:09:11.767794 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c444bf898-b6z6t"] Dec 05 10:09:11 crc kubenswrapper[4815]: I1205 10:09:11.993033 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c444bf898-b6z6t" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon-log" containerID="cri-o://d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b" gracePeriod=30 Dec 05 10:09:11 crc kubenswrapper[4815]: I1205 10:09:11.993096 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-c444bf898-b6z6t" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon" containerID="cri-o://8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa" gracePeriod=30 Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.264878 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.403449 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-dns-svc\") pod \"a06a9288-ec09-4ff0-8a21-c0a131880da8\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.403622 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m222v\" (UniqueName: \"kubernetes.io/projected/a06a9288-ec09-4ff0-8a21-c0a131880da8-kube-api-access-m222v\") pod \"a06a9288-ec09-4ff0-8a21-c0a131880da8\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.403666 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-config\") pod \"a06a9288-ec09-4ff0-8a21-c0a131880da8\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.403711 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-openstack-edpm-ipam\") pod \"a06a9288-ec09-4ff0-8a21-c0a131880da8\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.403754 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-nb\") pod \"a06a9288-ec09-4ff0-8a21-c0a131880da8\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.403773 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-sb\") pod \"a06a9288-ec09-4ff0-8a21-c0a131880da8\" (UID: \"a06a9288-ec09-4ff0-8a21-c0a131880da8\") " Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.411683 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a06a9288-ec09-4ff0-8a21-c0a131880da8-kube-api-access-m222v" (OuterVolumeSpecName: "kube-api-access-m222v") pod "a06a9288-ec09-4ff0-8a21-c0a131880da8" (UID: "a06a9288-ec09-4ff0-8a21-c0a131880da8"). InnerVolumeSpecName "kube-api-access-m222v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.507786 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m222v\" (UniqueName: \"kubernetes.io/projected/a06a9288-ec09-4ff0-8a21-c0a131880da8-kube-api-access-m222v\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.517024 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a06a9288-ec09-4ff0-8a21-c0a131880da8" (UID: "a06a9288-ec09-4ff0-8a21-c0a131880da8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.520208 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a06a9288-ec09-4ff0-8a21-c0a131880da8" (UID: "a06a9288-ec09-4ff0-8a21-c0a131880da8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.542045 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-config" (OuterVolumeSpecName: "config") pod "a06a9288-ec09-4ff0-8a21-c0a131880da8" (UID: "a06a9288-ec09-4ff0-8a21-c0a131880da8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.549780 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "a06a9288-ec09-4ff0-8a21-c0a131880da8" (UID: "a06a9288-ec09-4ff0-8a21-c0a131880da8"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.553791 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a06a9288-ec09-4ff0-8a21-c0a131880da8" (UID: "a06a9288-ec09-4ff0-8a21-c0a131880da8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.611374 4815 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.611407 4815 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-config\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.611419 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.611430 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:13 crc kubenswrapper[4815]: I1205 10:09:13.611440 4815 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a06a9288-ec09-4ff0-8a21-c0a131880da8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.027214 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"944629dc-36f8-4784-ac9c-7dea31445d5a","Type":"ContainerStarted","Data":"bd36b0e1df90a8d2c028d6bc96cfcdce4888106d1cc37c9607d8c87b7a6ad5c4"} Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.037247 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" event={"ID":"a06a9288-ec09-4ff0-8a21-c0a131880da8","Type":"ContainerDied","Data":"233da146f2f2de53321c030bb98a228cc959c0582eda3cab3a1b376175f58658"} Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.037299 4815 scope.go:117] "RemoveContainer" containerID="ee9878b701d75509d14856c88dd79be3eef3919f11c76c7f390446563bd86c7e" Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.037321 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-54fcc757dc-kbs2f" Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.075534 4815 scope.go:117] "RemoveContainer" containerID="28f263330a6907540525a3495bdb79ddbb3e0ea4b51331dd2dfc415185e384c2" Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.079155 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.079414 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="ceilometer-central-agent" containerID="cri-o://a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d" gracePeriod=30 Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.079566 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="proxy-httpd" containerID="cri-o://8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540" gracePeriod=30 Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.079610 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="sg-core" containerID="cri-o://69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4" gracePeriod=30 Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.079647 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="ceilometer-notification-agent" containerID="cri-o://a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0" gracePeriod=30 Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.099924 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-54fcc757dc-kbs2f"] Dec 05 10:09:14 crc kubenswrapper[4815]: I1205 10:09:14.114208 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-54fcc757dc-kbs2f"] Dec 05 10:09:15 crc kubenswrapper[4815]: I1205 10:09:15.047252 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"944629dc-36f8-4784-ac9c-7dea31445d5a","Type":"ContainerStarted","Data":"ba532672c9b6834118c19bc88aa23b8bc95d3519cffb43d166ff8e7126725751"} Dec 05 10:09:15 crc kubenswrapper[4815]: I1205 10:09:15.052415 4815 generic.go:334] "Generic (PLEG): container finished" podID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerID="8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540" exitCode=0 Dec 05 10:09:15 crc kubenswrapper[4815]: I1205 10:09:15.052443 4815 generic.go:334] "Generic (PLEG): container finished" podID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerID="69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4" exitCode=2 Dec 05 10:09:15 crc kubenswrapper[4815]: I1205 10:09:15.052450 4815 generic.go:334] "Generic (PLEG): container finished" podID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerID="a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d" exitCode=0 Dec 05 10:09:15 crc kubenswrapper[4815]: I1205 10:09:15.052454 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561c4cfb-f0f8-4f00-98f7-f4690d343047","Type":"ContainerDied","Data":"8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540"} Dec 05 10:09:15 crc kubenswrapper[4815]: I1205 10:09:15.052536 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561c4cfb-f0f8-4f00-98f7-f4690d343047","Type":"ContainerDied","Data":"69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4"} Dec 05 10:09:15 crc kubenswrapper[4815]: I1205 10:09:15.052554 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561c4cfb-f0f8-4f00-98f7-f4690d343047","Type":"ContainerDied","Data":"a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d"} Dec 05 10:09:15 crc kubenswrapper[4815]: I1205 10:09:15.068781 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=4.450414496 podStartE2EDuration="17.068764575s" podCreationTimestamp="2025-12-05 10:08:58 +0000 UTC" firstStartedPulling="2025-12-05 10:09:00.185330571 +0000 UTC m=+3739.063937408" lastFinishedPulling="2025-12-05 10:09:12.80368065 +0000 UTC m=+3751.682287487" observedRunningTime="2025-12-05 10:09:15.065415723 +0000 UTC m=+3753.944022560" watchObservedRunningTime="2025-12-05 10:09:15.068764575 +0000 UTC m=+3753.947371412" Dec 05 10:09:15 crc kubenswrapper[4815]: I1205 10:09:15.432284 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a06a9288-ec09-4ff0-8a21-c0a131880da8" path="/var/lib/kubelet/pods/a06a9288-ec09-4ff0-8a21-c0a131880da8/volumes" Dec 05 10:09:16 crc kubenswrapper[4815]: I1205 10:09:16.066906 4815 generic.go:334] "Generic (PLEG): container finished" podID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerID="8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa" exitCode=0 Dec 05 10:09:16 crc kubenswrapper[4815]: I1205 10:09:16.066991 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c444bf898-b6z6t" event={"ID":"819b83db-6f8e-4b8d-81d7-719760d1ff6b","Type":"ContainerDied","Data":"8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa"} Dec 05 10:09:16 crc kubenswrapper[4815]: I1205 10:09:16.485683 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c444bf898-b6z6t" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.852691 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.930601 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9rp7\" (UniqueName: \"kubernetes.io/projected/561c4cfb-f0f8-4f00-98f7-f4690d343047-kube-api-access-s9rp7\") pod \"561c4cfb-f0f8-4f00-98f7-f4690d343047\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.931870 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-run-httpd\") pod \"561c4cfb-f0f8-4f00-98f7-f4690d343047\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.931912 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-config-data\") pod \"561c4cfb-f0f8-4f00-98f7-f4690d343047\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.931936 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-combined-ca-bundle\") pod \"561c4cfb-f0f8-4f00-98f7-f4690d343047\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.932019 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-sg-core-conf-yaml\") pod \"561c4cfb-f0f8-4f00-98f7-f4690d343047\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.932069 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-scripts\") pod \"561c4cfb-f0f8-4f00-98f7-f4690d343047\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.932106 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-ceilometer-tls-certs\") pod \"561c4cfb-f0f8-4f00-98f7-f4690d343047\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.932153 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-log-httpd\") pod \"561c4cfb-f0f8-4f00-98f7-f4690d343047\" (UID: \"561c4cfb-f0f8-4f00-98f7-f4690d343047\") " Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.932184 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "561c4cfb-f0f8-4f00-98f7-f4690d343047" (UID: "561c4cfb-f0f8-4f00-98f7-f4690d343047"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.932542 4815 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.932916 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "561c4cfb-f0f8-4f00-98f7-f4690d343047" (UID: "561c4cfb-f0f8-4f00-98f7-f4690d343047"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.944442 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561c4cfb-f0f8-4f00-98f7-f4690d343047-kube-api-access-s9rp7" (OuterVolumeSpecName: "kube-api-access-s9rp7") pod "561c4cfb-f0f8-4f00-98f7-f4690d343047" (UID: "561c4cfb-f0f8-4f00-98f7-f4690d343047"). InnerVolumeSpecName "kube-api-access-s9rp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:09:17 crc kubenswrapper[4815]: I1205 10:09:17.944999 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-scripts" (OuterVolumeSpecName: "scripts") pod "561c4cfb-f0f8-4f00-98f7-f4690d343047" (UID: "561c4cfb-f0f8-4f00-98f7-f4690d343047"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.035716 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.035743 4815 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/561c4cfb-f0f8-4f00-98f7-f4690d343047-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.035752 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9rp7\" (UniqueName: \"kubernetes.io/projected/561c4cfb-f0f8-4f00-98f7-f4690d343047-kube-api-access-s9rp7\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.078177 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "561c4cfb-f0f8-4f00-98f7-f4690d343047" (UID: "561c4cfb-f0f8-4f00-98f7-f4690d343047"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.087942 4815 generic.go:334] "Generic (PLEG): container finished" podID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerID="a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0" exitCode=0 Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.087985 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561c4cfb-f0f8-4f00-98f7-f4690d343047","Type":"ContainerDied","Data":"a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0"} Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.088014 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"561c4cfb-f0f8-4f00-98f7-f4690d343047","Type":"ContainerDied","Data":"bd99c9a2c987e5bdb7ee66c534b8c9743dc4a4cb486d08b575ab308407005379"} Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.088029 4815 scope.go:117] "RemoveContainer" containerID="8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.088152 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.131907 4815 scope.go:117] "RemoveContainer" containerID="69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.137886 4815 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.138410 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "561c4cfb-f0f8-4f00-98f7-f4690d343047" (UID: "561c4cfb-f0f8-4f00-98f7-f4690d343047"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.154390 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "561c4cfb-f0f8-4f00-98f7-f4690d343047" (UID: "561c4cfb-f0f8-4f00-98f7-f4690d343047"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.154600 4815 scope.go:117] "RemoveContainer" containerID="a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.180779 4815 scope.go:117] "RemoveContainer" containerID="a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.190627 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-config-data" (OuterVolumeSpecName: "config-data") pod "561c4cfb-f0f8-4f00-98f7-f4690d343047" (UID: "561c4cfb-f0f8-4f00-98f7-f4690d343047"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.206919 4815 scope.go:117] "RemoveContainer" containerID="8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540" Dec 05 10:09:18 crc kubenswrapper[4815]: E1205 10:09:18.207364 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540\": container with ID starting with 8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540 not found: ID does not exist" containerID="8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.207397 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540"} err="failed to get container status \"8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540\": rpc error: code = NotFound desc = could not find container \"8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540\": container with ID starting with 8ebcabdf5b2068042324840b5ccdb927487bc7fe66ef3aa754905e6ddd687540 not found: ID does not exist" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.207417 4815 scope.go:117] "RemoveContainer" containerID="69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4" Dec 05 10:09:18 crc kubenswrapper[4815]: E1205 10:09:18.208641 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4\": container with ID starting with 69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4 not found: ID does not exist" containerID="69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.208664 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4"} err="failed to get container status \"69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4\": rpc error: code = NotFound desc = could not find container \"69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4\": container with ID starting with 69a9485791b1338a4a51cdfb70addb395f3ee1ca2ba2a4a900648434620d8de4 not found: ID does not exist" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.208678 4815 scope.go:117] "RemoveContainer" containerID="a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0" Dec 05 10:09:18 crc kubenswrapper[4815]: E1205 10:09:18.208878 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0\": container with ID starting with a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0 not found: ID does not exist" containerID="a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.208914 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0"} err="failed to get container status \"a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0\": rpc error: code = NotFound desc = could not find container \"a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0\": container with ID starting with a446ab5e7ff4eaa8e603a84b09c0ff4d4a39e1fe7b8886966a23b8bc87b105c0 not found: ID does not exist" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.208933 4815 scope.go:117] "RemoveContainer" containerID="a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d" Dec 05 10:09:18 crc kubenswrapper[4815]: E1205 10:09:18.209212 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d\": container with ID starting with a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d not found: ID does not exist" containerID="a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.209235 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d"} err="failed to get container status \"a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d\": rpc error: code = NotFound desc = could not find container \"a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d\": container with ID starting with a866aa18e9c12648135c327393aa4577798273295baf1c3959b064058ca6e83d not found: ID does not exist" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.240518 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.240552 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.240565 4815 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/561c4cfb-f0f8-4f00-98f7-f4690d343047-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.427694 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.434646 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.455335 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 05 10:09:18 crc kubenswrapper[4815]: E1205 10:09:18.456120 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="ceilometer-central-agent" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.456233 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="ceilometer-central-agent" Dec 05 10:09:18 crc kubenswrapper[4815]: E1205 10:09:18.456352 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="proxy-httpd" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.456431 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="proxy-httpd" Dec 05 10:09:18 crc kubenswrapper[4815]: E1205 10:09:18.456550 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a06a9288-ec09-4ff0-8a21-c0a131880da8" containerName="dnsmasq-dns" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.456660 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a06a9288-ec09-4ff0-8a21-c0a131880da8" containerName="dnsmasq-dns" Dec 05 10:09:18 crc kubenswrapper[4815]: E1205 10:09:18.456757 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a06a9288-ec09-4ff0-8a21-c0a131880da8" containerName="init" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.456833 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="a06a9288-ec09-4ff0-8a21-c0a131880da8" containerName="init" Dec 05 10:09:18 crc kubenswrapper[4815]: E1205 10:09:18.456938 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="sg-core" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.457019 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="sg-core" Dec 05 10:09:18 crc kubenswrapper[4815]: E1205 10:09:18.457124 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="ceilometer-notification-agent" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.457200 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="ceilometer-notification-agent" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.457621 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="a06a9288-ec09-4ff0-8a21-c0a131880da8" containerName="dnsmasq-dns" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.457747 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="ceilometer-central-agent" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.457829 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="sg-core" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.457933 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="proxy-httpd" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.458026 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" containerName="ceilometer-notification-agent" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.460291 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.468092 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.468337 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.468419 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.472012 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.647960 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-scripts\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.648747 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.648973 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qx44j\" (UniqueName: \"kubernetes.io/projected/078b665f-5e54-4d83-966b-3684d06ea320-kube-api-access-qx44j\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.649221 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/078b665f-5e54-4d83-966b-3684d06ea320-log-httpd\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.649439 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.649825 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/078b665f-5e54-4d83-966b-3684d06ea320-run-httpd\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.652142 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.652219 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-config-data\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.754192 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.754254 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-config-data\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.754294 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-scripts\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.754358 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.754382 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qx44j\" (UniqueName: \"kubernetes.io/projected/078b665f-5e54-4d83-966b-3684d06ea320-kube-api-access-qx44j\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.754410 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/078b665f-5e54-4d83-966b-3684d06ea320-log-httpd\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.754429 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.754474 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/078b665f-5e54-4d83-966b-3684d06ea320-run-httpd\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.755093 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/078b665f-5e54-4d83-966b-3684d06ea320-run-httpd\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.756549 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/078b665f-5e54-4d83-966b-3684d06ea320-log-httpd\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:18 crc kubenswrapper[4815]: I1205 10:09:18.892087 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 05 10:09:19 crc kubenswrapper[4815]: I1205 10:09:19.126543 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:19 crc kubenswrapper[4815]: I1205 10:09:19.126839 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:19 crc kubenswrapper[4815]: I1205 10:09:19.126992 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:19 crc kubenswrapper[4815]: I1205 10:09:19.127037 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-config-data\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:19 crc kubenswrapper[4815]: I1205 10:09:19.128046 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/078b665f-5e54-4d83-966b-3684d06ea320-scripts\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:19 crc kubenswrapper[4815]: I1205 10:09:19.128534 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qx44j\" (UniqueName: \"kubernetes.io/projected/078b665f-5e54-4d83-966b-3684d06ea320-kube-api-access-qx44j\") pod \"ceilometer-0\" (UID: \"078b665f-5e54-4d83-966b-3684d06ea320\") " pod="openstack/ceilometer-0" Dec 05 10:09:19 crc kubenswrapper[4815]: I1205 10:09:19.400614 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 05 10:09:19 crc kubenswrapper[4815]: I1205 10:09:19.430012 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561c4cfb-f0f8-4f00-98f7-f4690d343047" path="/var/lib/kubelet/pods/561c4cfb-f0f8-4f00-98f7-f4690d343047/volumes" Dec 05 10:09:19 crc kubenswrapper[4815]: I1205 10:09:19.869591 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 05 10:09:20 crc kubenswrapper[4815]: I1205 10:09:20.111598 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"078b665f-5e54-4d83-966b-3684d06ea320","Type":"ContainerStarted","Data":"b310186ab1be9abda3b40178d3f01882880e8fd4a406afe48b0807b4b03874ea"} Dec 05 10:09:20 crc kubenswrapper[4815]: I1205 10:09:20.587565 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 05 10:09:20 crc kubenswrapper[4815]: I1205 10:09:20.655073 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 10:09:21 crc kubenswrapper[4815]: I1205 10:09:21.123272 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="551a6e4e-9e8a-4aca-b569-966f6f562903" containerName="manila-scheduler" containerID="cri-o://5f00bf238719fb0032f9be6155c09e2a5b585fcf3cdedd6e29b08ed3e0fb3556" gracePeriod=30 Dec 05 10:09:21 crc kubenswrapper[4815]: I1205 10:09:21.123654 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"078b665f-5e54-4d83-966b-3684d06ea320","Type":"ContainerStarted","Data":"a5fdfe48c4d2a7d534d0e15d19cbe6bc21987585935e745c549f213c967eda4c"} Dec 05 10:09:21 crc kubenswrapper[4815]: I1205 10:09:21.124095 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="551a6e4e-9e8a-4aca-b569-966f6f562903" containerName="probe" containerID="cri-o://ba8ba55cd835b36ab2fa6a0c5df436127112db8452432ecc3007ad8a7ceb9e2b" gracePeriod=30 Dec 05 10:09:22 crc kubenswrapper[4815]: I1205 10:09:22.132690 4815 generic.go:334] "Generic (PLEG): container finished" podID="551a6e4e-9e8a-4aca-b569-966f6f562903" containerID="ba8ba55cd835b36ab2fa6a0c5df436127112db8452432ecc3007ad8a7ceb9e2b" exitCode=0 Dec 05 10:09:22 crc kubenswrapper[4815]: I1205 10:09:22.132765 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"551a6e4e-9e8a-4aca-b569-966f6f562903","Type":"ContainerDied","Data":"ba8ba55cd835b36ab2fa6a0c5df436127112db8452432ecc3007ad8a7ceb9e2b"} Dec 05 10:09:22 crc kubenswrapper[4815]: I1205 10:09:22.135097 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"078b665f-5e54-4d83-966b-3684d06ea320","Type":"ContainerStarted","Data":"3e24a3d33e03db57c2060a67571e79020e5adc2d2ecf410b67e756baaaefcb5a"} Dec 05 10:09:23 crc kubenswrapper[4815]: I1205 10:09:23.145234 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"078b665f-5e54-4d83-966b-3684d06ea320","Type":"ContainerStarted","Data":"097d2d10f5e82e02b24a9dc9f6777e67793086cebb6baa138bbe034c773a18b2"} Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.194142 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"078b665f-5e54-4d83-966b-3684d06ea320","Type":"ContainerStarted","Data":"1ddc68c88ae795953c1dd7c0a0cce21c779c0edabd971b25e2412ad0c754a996"} Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.194920 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.207471 4815 generic.go:334] "Generic (PLEG): container finished" podID="551a6e4e-9e8a-4aca-b569-966f6f562903" containerID="5f00bf238719fb0032f9be6155c09e2a5b585fcf3cdedd6e29b08ed3e0fb3556" exitCode=0 Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.207646 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"551a6e4e-9e8a-4aca-b569-966f6f562903","Type":"ContainerDied","Data":"5f00bf238719fb0032f9be6155c09e2a5b585fcf3cdedd6e29b08ed3e0fb3556"} Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.236072 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5560981270000003 podStartE2EDuration="7.236052254s" podCreationTimestamp="2025-12-05 10:09:18 +0000 UTC" firstStartedPulling="2025-12-05 10:09:19.877177946 +0000 UTC m=+3758.755784813" lastFinishedPulling="2025-12-05 10:09:24.557132103 +0000 UTC m=+3763.435738940" observedRunningTime="2025-12-05 10:09:25.23111925 +0000 UTC m=+3764.109726097" watchObservedRunningTime="2025-12-05 10:09:25.236052254 +0000 UTC m=+3764.114659101" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.319106 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.403193 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data-custom\") pod \"551a6e4e-9e8a-4aca-b569-966f6f562903\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.403355 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-scripts\") pod \"551a6e4e-9e8a-4aca-b569-966f6f562903\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.403390 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/551a6e4e-9e8a-4aca-b569-966f6f562903-etc-machine-id\") pod \"551a6e4e-9e8a-4aca-b569-966f6f562903\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.403470 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data\") pod \"551a6e4e-9e8a-4aca-b569-966f6f562903\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.403645 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/551a6e4e-9e8a-4aca-b569-966f6f562903-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "551a6e4e-9e8a-4aca-b569-966f6f562903" (UID: "551a6e4e-9e8a-4aca-b569-966f6f562903"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.403692 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-combined-ca-bundle\") pod \"551a6e4e-9e8a-4aca-b569-966f6f562903\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.403737 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s9vss\" (UniqueName: \"kubernetes.io/projected/551a6e4e-9e8a-4aca-b569-966f6f562903-kube-api-access-s9vss\") pod \"551a6e4e-9e8a-4aca-b569-966f6f562903\" (UID: \"551a6e4e-9e8a-4aca-b569-966f6f562903\") " Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.404218 4815 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/551a6e4e-9e8a-4aca-b569-966f6f562903-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.410865 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-scripts" (OuterVolumeSpecName: "scripts") pod "551a6e4e-9e8a-4aca-b569-966f6f562903" (UID: "551a6e4e-9e8a-4aca-b569-966f6f562903"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.413816 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/551a6e4e-9e8a-4aca-b569-966f6f562903-kube-api-access-s9vss" (OuterVolumeSpecName: "kube-api-access-s9vss") pod "551a6e4e-9e8a-4aca-b569-966f6f562903" (UID: "551a6e4e-9e8a-4aca-b569-966f6f562903"). InnerVolumeSpecName "kube-api-access-s9vss". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.419681 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "551a6e4e-9e8a-4aca-b569-966f6f562903" (UID: "551a6e4e-9e8a-4aca-b569-966f6f562903"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.420575 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:09:25 crc kubenswrapper[4815]: E1205 10:09:25.420953 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.479600 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "551a6e4e-9e8a-4aca-b569-966f6f562903" (UID: "551a6e4e-9e8a-4aca-b569-966f6f562903"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.506402 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.506691 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s9vss\" (UniqueName: \"kubernetes.io/projected/551a6e4e-9e8a-4aca-b569-966f6f562903-kube-api-access-s9vss\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.506780 4815 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.506908 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.533100 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data" (OuterVolumeSpecName: "config-data") pod "551a6e4e-9e8a-4aca-b569-966f6f562903" (UID: "551a6e4e-9e8a-4aca-b569-966f6f562903"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:25 crc kubenswrapper[4815]: I1205 10:09:25.616828 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551a6e4e-9e8a-4aca-b569-966f6f562903-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.218318 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"551a6e4e-9e8a-4aca-b569-966f6f562903","Type":"ContainerDied","Data":"97a41535c7b9f258138807d847e89332e3b78b2da7139318526d3417acef3e45"} Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.218354 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.218391 4815 scope.go:117] "RemoveContainer" containerID="ba8ba55cd835b36ab2fa6a0c5df436127112db8452432ecc3007ad8a7ceb9e2b" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.283960 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.295512 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.307939 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 10:09:26 crc kubenswrapper[4815]: E1205 10:09:26.311073 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551a6e4e-9e8a-4aca-b569-966f6f562903" containerName="probe" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.311344 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="551a6e4e-9e8a-4aca-b569-966f6f562903" containerName="probe" Dec 05 10:09:26 crc kubenswrapper[4815]: E1205 10:09:26.311505 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551a6e4e-9e8a-4aca-b569-966f6f562903" containerName="manila-scheduler" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.311621 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="551a6e4e-9e8a-4aca-b569-966f6f562903" containerName="manila-scheduler" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.312134 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="551a6e4e-9e8a-4aca-b569-966f6f562903" containerName="probe" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.312277 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="551a6e4e-9e8a-4aca-b569-966f6f562903" containerName="manila-scheduler" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.314475 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.316992 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.322161 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.432416 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.433159 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-scripts\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.433352 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zdxl\" (UniqueName: \"kubernetes.io/projected/bed98f1a-2e43-47cc-82af-ca5c1a833995-kube-api-access-8zdxl\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.433586 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bed98f1a-2e43-47cc-82af-ca5c1a833995-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.433778 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-config-data\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.433934 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.456070 4815 scope.go:117] "RemoveContainer" containerID="5f00bf238719fb0032f9be6155c09e2a5b585fcf3cdedd6e29b08ed3e0fb3556" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.485199 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c444bf898-b6z6t" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.536078 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.536142 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-scripts\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.536217 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zdxl\" (UniqueName: \"kubernetes.io/projected/bed98f1a-2e43-47cc-82af-ca5c1a833995-kube-api-access-8zdxl\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.536279 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bed98f1a-2e43-47cc-82af-ca5c1a833995-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.536352 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-config-data\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.536414 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.537438 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/bed98f1a-2e43-47cc-82af-ca5c1a833995-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.547208 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.548499 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-scripts\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.550012 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.551867 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bed98f1a-2e43-47cc-82af-ca5c1a833995-config-data\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.560356 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zdxl\" (UniqueName: \"kubernetes.io/projected/bed98f1a-2e43-47cc-82af-ca5c1a833995-kube-api-access-8zdxl\") pod \"manila-scheduler-0\" (UID: \"bed98f1a-2e43-47cc-82af-ca5c1a833995\") " pod="openstack/manila-scheduler-0" Dec 05 10:09:26 crc kubenswrapper[4815]: I1205 10:09:26.638831 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 05 10:09:27 crc kubenswrapper[4815]: I1205 10:09:27.218823 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 05 10:09:27 crc kubenswrapper[4815]: W1205 10:09:27.226716 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbed98f1a_2e43_47cc_82af_ca5c1a833995.slice/crio-4d0fc15d832c8e72a69a886cb3b05abc0c19ea0418772b71d000b11e4763befc WatchSource:0}: Error finding container 4d0fc15d832c8e72a69a886cb3b05abc0c19ea0418772b71d000b11e4763befc: Status 404 returned error can't find the container with id 4d0fc15d832c8e72a69a886cb3b05abc0c19ea0418772b71d000b11e4763befc Dec 05 10:09:27 crc kubenswrapper[4815]: I1205 10:09:27.430773 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="551a6e4e-9e8a-4aca-b569-966f6f562903" path="/var/lib/kubelet/pods/551a6e4e-9e8a-4aca-b569-966f6f562903/volumes" Dec 05 10:09:27 crc kubenswrapper[4815]: I1205 10:09:27.713352 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 05 10:09:28 crc kubenswrapper[4815]: I1205 10:09:28.250402 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"bed98f1a-2e43-47cc-82af-ca5c1a833995","Type":"ContainerStarted","Data":"4d0fc15d832c8e72a69a886cb3b05abc0c19ea0418772b71d000b11e4763befc"} Dec 05 10:09:29 crc kubenswrapper[4815]: I1205 10:09:29.263682 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"bed98f1a-2e43-47cc-82af-ca5c1a833995","Type":"ContainerStarted","Data":"2f475aeda9b4f5fd11f7c44d0f363ad989db071a853411fdc679cf833e0e4b22"} Dec 05 10:09:29 crc kubenswrapper[4815]: I1205 10:09:29.263959 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"bed98f1a-2e43-47cc-82af-ca5c1a833995","Type":"ContainerStarted","Data":"21994c69d650376ef059d37450ec5eb54002f2f595eec04db340eccda24ecb74"} Dec 05 10:09:30 crc kubenswrapper[4815]: I1205 10:09:30.640787 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 05 10:09:30 crc kubenswrapper[4815]: I1205 10:09:30.680132 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=4.680106996 podStartE2EDuration="4.680106996s" podCreationTimestamp="2025-12-05 10:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:09:29.295079711 +0000 UTC m=+3768.173686548" watchObservedRunningTime="2025-12-05 10:09:30.680106996 +0000 UTC m=+3769.558713843" Dec 05 10:09:30 crc kubenswrapper[4815]: I1205 10:09:30.737354 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 10:09:31 crc kubenswrapper[4815]: I1205 10:09:31.281945 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="944629dc-36f8-4784-ac9c-7dea31445d5a" containerName="manila-share" containerID="cri-o://bd36b0e1df90a8d2c028d6bc96cfcdce4888106d1cc37c9607d8c87b7a6ad5c4" gracePeriod=30 Dec 05 10:09:31 crc kubenswrapper[4815]: I1205 10:09:31.282692 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="944629dc-36f8-4784-ac9c-7dea31445d5a" containerName="probe" containerID="cri-o://ba532672c9b6834118c19bc88aa23b8bc95d3519cffb43d166ff8e7126725751" gracePeriod=30 Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.294860 4815 generic.go:334] "Generic (PLEG): container finished" podID="944629dc-36f8-4784-ac9c-7dea31445d5a" containerID="ba532672c9b6834118c19bc88aa23b8bc95d3519cffb43d166ff8e7126725751" exitCode=0 Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.295075 4815 generic.go:334] "Generic (PLEG): container finished" podID="944629dc-36f8-4784-ac9c-7dea31445d5a" containerID="bd36b0e1df90a8d2c028d6bc96cfcdce4888106d1cc37c9607d8c87b7a6ad5c4" exitCode=1 Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.294948 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"944629dc-36f8-4784-ac9c-7dea31445d5a","Type":"ContainerDied","Data":"ba532672c9b6834118c19bc88aa23b8bc95d3519cffb43d166ff8e7126725751"} Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.295112 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"944629dc-36f8-4784-ac9c-7dea31445d5a","Type":"ContainerDied","Data":"bd36b0e1df90a8d2c028d6bc96cfcdce4888106d1cc37c9607d8c87b7a6ad5c4"} Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.433072 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.598784 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-ceph\") pod \"944629dc-36f8-4784-ac9c-7dea31445d5a\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.598855 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-scripts\") pod \"944629dc-36f8-4784-ac9c-7dea31445d5a\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.598901 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-var-lib-manila\") pod \"944629dc-36f8-4784-ac9c-7dea31445d5a\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.598921 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-etc-machine-id\") pod \"944629dc-36f8-4784-ac9c-7dea31445d5a\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.598952 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data-custom\") pod \"944629dc-36f8-4784-ac9c-7dea31445d5a\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.599052 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "944629dc-36f8-4784-ac9c-7dea31445d5a" (UID: "944629dc-36f8-4784-ac9c-7dea31445d5a"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.599067 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data\") pod \"944629dc-36f8-4784-ac9c-7dea31445d5a\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.599157 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-combined-ca-bundle\") pod \"944629dc-36f8-4784-ac9c-7dea31445d5a\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.599208 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb7bt\" (UniqueName: \"kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-kube-api-access-xb7bt\") pod \"944629dc-36f8-4784-ac9c-7dea31445d5a\" (UID: \"944629dc-36f8-4784-ac9c-7dea31445d5a\") " Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.600229 4815 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-var-lib-manila\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.602187 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "944629dc-36f8-4784-ac9c-7dea31445d5a" (UID: "944629dc-36f8-4784-ac9c-7dea31445d5a"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.606058 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-scripts" (OuterVolumeSpecName: "scripts") pod "944629dc-36f8-4784-ac9c-7dea31445d5a" (UID: "944629dc-36f8-4784-ac9c-7dea31445d5a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.606226 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-ceph" (OuterVolumeSpecName: "ceph") pod "944629dc-36f8-4784-ac9c-7dea31445d5a" (UID: "944629dc-36f8-4784-ac9c-7dea31445d5a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.607326 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "944629dc-36f8-4784-ac9c-7dea31445d5a" (UID: "944629dc-36f8-4784-ac9c-7dea31445d5a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.608601 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-kube-api-access-xb7bt" (OuterVolumeSpecName: "kube-api-access-xb7bt") pod "944629dc-36f8-4784-ac9c-7dea31445d5a" (UID: "944629dc-36f8-4784-ac9c-7dea31445d5a"). InnerVolumeSpecName "kube-api-access-xb7bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.671733 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "944629dc-36f8-4784-ac9c-7dea31445d5a" (UID: "944629dc-36f8-4784-ac9c-7dea31445d5a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.701806 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.701838 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb7bt\" (UniqueName: \"kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-kube-api-access-xb7bt\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.701849 4815 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/944629dc-36f8-4784-ac9c-7dea31445d5a-ceph\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.701858 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.701866 4815 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/944629dc-36f8-4784-ac9c-7dea31445d5a-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.701874 4815 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.718023 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data" (OuterVolumeSpecName: "config-data") pod "944629dc-36f8-4784-ac9c-7dea31445d5a" (UID: "944629dc-36f8-4784-ac9c-7dea31445d5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:32 crc kubenswrapper[4815]: I1205 10:09:32.803452 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/944629dc-36f8-4784-ac9c-7dea31445d5a-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.306226 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"944629dc-36f8-4784-ac9c-7dea31445d5a","Type":"ContainerDied","Data":"d234426f46b59747c9df7c1678726b4079d5c7e2ad7d94d913cf5eaeb22c20d2"} Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.307621 4815 scope.go:117] "RemoveContainer" containerID="ba532672c9b6834118c19bc88aa23b8bc95d3519cffb43d166ff8e7126725751" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.306276 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.340865 4815 scope.go:117] "RemoveContainer" containerID="bd36b0e1df90a8d2c028d6bc96cfcdce4888106d1cc37c9607d8c87b7a6ad5c4" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.350265 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.359529 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.397074 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 10:09:33 crc kubenswrapper[4815]: E1205 10:09:33.397481 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="944629dc-36f8-4784-ac9c-7dea31445d5a" containerName="probe" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.397512 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="944629dc-36f8-4784-ac9c-7dea31445d5a" containerName="probe" Dec 05 10:09:33 crc kubenswrapper[4815]: E1205 10:09:33.397537 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="944629dc-36f8-4784-ac9c-7dea31445d5a" containerName="manila-share" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.397544 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="944629dc-36f8-4784-ac9c-7dea31445d5a" containerName="manila-share" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.397727 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="944629dc-36f8-4784-ac9c-7dea31445d5a" containerName="probe" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.397750 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="944629dc-36f8-4784-ac9c-7dea31445d5a" containerName="manila-share" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.398886 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.402747 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.408712 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.437727 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="944629dc-36f8-4784-ac9c-7dea31445d5a" path="/var/lib/kubelet/pods/944629dc-36f8-4784-ac9c-7dea31445d5a/volumes" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.522513 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.522659 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.522769 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-ceph\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.522911 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.522977 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.523029 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-config-data\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.523089 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-scripts\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.523109 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tv6n\" (UniqueName: \"kubernetes.io/projected/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-kube-api-access-9tv6n\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.624524 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-scripts\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.624771 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tv6n\" (UniqueName: \"kubernetes.io/projected/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-kube-api-access-9tv6n\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.624956 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.625618 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.625767 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-ceph\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.625980 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.626314 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.626374 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.626419 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.626447 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-config-data\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.629566 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-scripts\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.629967 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-ceph\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.630263 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.631998 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.636017 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-config-data\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.655835 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tv6n\" (UniqueName: \"kubernetes.io/projected/7d69d7b9-98dc-4dff-96bc-e562b8bc4b98-kube-api-access-9tv6n\") pod \"manila-share-share1-0\" (UID: \"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98\") " pod="openstack/manila-share-share1-0" Dec 05 10:09:33 crc kubenswrapper[4815]: I1205 10:09:33.727513 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 05 10:09:34 crc kubenswrapper[4815]: W1205 10:09:34.397920 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d69d7b9_98dc_4dff_96bc_e562b8bc4b98.slice/crio-de2c08a9abc11b87d6546ed5fec2fac5c37b5aa27e525a1cc6e99c3c3ce275eb WatchSource:0}: Error finding container de2c08a9abc11b87d6546ed5fec2fac5c37b5aa27e525a1cc6e99c3c3ce275eb: Status 404 returned error can't find the container with id de2c08a9abc11b87d6546ed5fec2fac5c37b5aa27e525a1cc6e99c3c3ce275eb Dec 05 10:09:34 crc kubenswrapper[4815]: I1205 10:09:34.398997 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 05 10:09:35 crc kubenswrapper[4815]: I1205 10:09:35.328950 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98","Type":"ContainerStarted","Data":"b2d111f4b840fb2ed24a7d7fb04e8819e42890811961db999f65695fcb02ad9e"} Dec 05 10:09:35 crc kubenswrapper[4815]: I1205 10:09:35.329428 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98","Type":"ContainerStarted","Data":"069c8f0e81904ed495cbbf36cd9a1b3986a4dcae6557096ff50c55adb52a2eb6"} Dec 05 10:09:35 crc kubenswrapper[4815]: I1205 10:09:35.329535 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"7d69d7b9-98dc-4dff-96bc-e562b8bc4b98","Type":"ContainerStarted","Data":"de2c08a9abc11b87d6546ed5fec2fac5c37b5aa27e525a1cc6e99c3c3ce275eb"} Dec 05 10:09:35 crc kubenswrapper[4815]: I1205 10:09:35.356214 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=2.356183187 podStartE2EDuration="2.356183187s" podCreationTimestamp="2025-12-05 10:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:09:35.35479983 +0000 UTC m=+3774.233406677" watchObservedRunningTime="2025-12-05 10:09:35.356183187 +0000 UTC m=+3774.234790024" Dec 05 10:09:36 crc kubenswrapper[4815]: I1205 10:09:36.420872 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:09:36 crc kubenswrapper[4815]: E1205 10:09:36.421326 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:09:36 crc kubenswrapper[4815]: I1205 10:09:36.485521 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-c444bf898-b6z6t" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.247:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.247:8443: connect: connection refused" Dec 05 10:09:36 crc kubenswrapper[4815]: I1205 10:09:36.485661 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:09:36 crc kubenswrapper[4815]: I1205 10:09:36.639078 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.407066 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.422458 4815 generic.go:334] "Generic (PLEG): container finished" podID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerID="d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b" exitCode=137 Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.422552 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c444bf898-b6z6t" event={"ID":"819b83db-6f8e-4b8d-81d7-719760d1ff6b","Type":"ContainerDied","Data":"d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b"} Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.422587 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-c444bf898-b6z6t" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.422608 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-c444bf898-b6z6t" event={"ID":"819b83db-6f8e-4b8d-81d7-719760d1ff6b","Type":"ContainerDied","Data":"928391e4b689d1a0e503c4b5e20313b29717362c6d33205338f6b72cfdbf0292"} Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.422627 4815 scope.go:117] "RemoveContainer" containerID="8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.556000 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-combined-ca-bundle\") pod \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.556312 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-scripts\") pod \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.556340 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-config-data\") pod \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.556804 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-tls-certs\") pod \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.556874 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m2lg6\" (UniqueName: \"kubernetes.io/projected/819b83db-6f8e-4b8d-81d7-719760d1ff6b-kube-api-access-m2lg6\") pod \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.556935 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-secret-key\") pod \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.557995 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/819b83db-6f8e-4b8d-81d7-719760d1ff6b-logs\") pod \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\" (UID: \"819b83db-6f8e-4b8d-81d7-719760d1ff6b\") " Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.558696 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/819b83db-6f8e-4b8d-81d7-719760d1ff6b-logs" (OuterVolumeSpecName: "logs") pod "819b83db-6f8e-4b8d-81d7-719760d1ff6b" (UID: "819b83db-6f8e-4b8d-81d7-719760d1ff6b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.559032 4815 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/819b83db-6f8e-4b8d-81d7-719760d1ff6b-logs\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.562755 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/819b83db-6f8e-4b8d-81d7-719760d1ff6b-kube-api-access-m2lg6" (OuterVolumeSpecName: "kube-api-access-m2lg6") pod "819b83db-6f8e-4b8d-81d7-719760d1ff6b" (UID: "819b83db-6f8e-4b8d-81d7-719760d1ff6b"). InnerVolumeSpecName "kube-api-access-m2lg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.566619 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "819b83db-6f8e-4b8d-81d7-719760d1ff6b" (UID: "819b83db-6f8e-4b8d-81d7-719760d1ff6b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.586669 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-config-data" (OuterVolumeSpecName: "config-data") pod "819b83db-6f8e-4b8d-81d7-719760d1ff6b" (UID: "819b83db-6f8e-4b8d-81d7-719760d1ff6b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.587245 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-scripts" (OuterVolumeSpecName: "scripts") pod "819b83db-6f8e-4b8d-81d7-719760d1ff6b" (UID: "819b83db-6f8e-4b8d-81d7-719760d1ff6b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.595546 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "819b83db-6f8e-4b8d-81d7-719760d1ff6b" (UID: "819b83db-6f8e-4b8d-81d7-719760d1ff6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.618670 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "819b83db-6f8e-4b8d-81d7-719760d1ff6b" (UID: "819b83db-6f8e-4b8d-81d7-719760d1ff6b"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.629817 4815 scope.go:117] "RemoveContainer" containerID="d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.660916 4815 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.660953 4815 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.660964 4815 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-scripts\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.660973 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/819b83db-6f8e-4b8d-81d7-719760d1ff6b-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.660981 4815 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/819b83db-6f8e-4b8d-81d7-719760d1ff6b-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.660991 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m2lg6\" (UniqueName: \"kubernetes.io/projected/819b83db-6f8e-4b8d-81d7-719760d1ff6b-kube-api-access-m2lg6\") on node \"crc\" DevicePath \"\"" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.662517 4815 scope.go:117] "RemoveContainer" containerID="8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa" Dec 05 10:09:42 crc kubenswrapper[4815]: E1205 10:09:42.662916 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa\": container with ID starting with 8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa not found: ID does not exist" containerID="8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.662973 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa"} err="failed to get container status \"8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa\": rpc error: code = NotFound desc = could not find container \"8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa\": container with ID starting with 8f7d3dfd5721046351246b0ddcf23646009628d366729a82eaa8e0987ab367fa not found: ID does not exist" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.663004 4815 scope.go:117] "RemoveContainer" containerID="d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b" Dec 05 10:09:42 crc kubenswrapper[4815]: E1205 10:09:42.663255 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b\": container with ID starting with d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b not found: ID does not exist" containerID="d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.663339 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b"} err="failed to get container status \"d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b\": rpc error: code = NotFound desc = could not find container \"d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b\": container with ID starting with d0d60df06dcc9e7814db33a72301920222c22a541ec8ed6b93cd262ffa30961b not found: ID does not exist" Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.757544 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-c444bf898-b6z6t"] Dec 05 10:09:42 crc kubenswrapper[4815]: I1205 10:09:42.767311 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-c444bf898-b6z6t"] Dec 05 10:09:43 crc kubenswrapper[4815]: I1205 10:09:43.457862 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" path="/var/lib/kubelet/pods/819b83db-6f8e-4b8d-81d7-719760d1ff6b/volumes" Dec 05 10:09:43 crc kubenswrapper[4815]: I1205 10:09:43.729156 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 05 10:09:48 crc kubenswrapper[4815]: I1205 10:09:48.282876 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 05 10:09:49 crc kubenswrapper[4815]: I1205 10:09:49.414709 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 05 10:09:49 crc kubenswrapper[4815]: I1205 10:09:49.419841 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:09:49 crc kubenswrapper[4815]: E1205 10:09:49.420281 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:09:55 crc kubenswrapper[4815]: I1205 10:09:55.277682 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 05 10:10:03 crc kubenswrapper[4815]: I1205 10:10:03.419382 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:10:03 crc kubenswrapper[4815]: E1205 10:10:03.420152 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:10:15 crc kubenswrapper[4815]: I1205 10:10:15.421647 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:10:15 crc kubenswrapper[4815]: E1205 10:10:15.423629 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:10:30 crc kubenswrapper[4815]: I1205 10:10:30.419353 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:10:30 crc kubenswrapper[4815]: E1205 10:10:30.420387 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:10:43 crc kubenswrapper[4815]: I1205 10:10:43.421232 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:10:43 crc kubenswrapper[4815]: E1205 10:10:43.421919 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:10:55 crc kubenswrapper[4815]: I1205 10:10:55.419264 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:10:55 crc kubenswrapper[4815]: E1205 10:10:55.420157 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.277675 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 05 10:11:01 crc kubenswrapper[4815]: E1205 10:11:01.279800 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon-log" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.281193 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon-log" Dec 05 10:11:01 crc kubenswrapper[4815]: E1205 10:11:01.281328 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.281410 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.281872 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon-log" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.281995 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="819b83db-6f8e-4b8d-81d7-719760d1ff6b" containerName="horizon" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.287624 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.304517 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.305023 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.305587 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.305905 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.305985 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-r7768" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.431922 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.431975 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.432016 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-config-data\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.432215 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45lb8\" (UniqueName: \"kubernetes.io/projected/b0ec9924-8223-4a07-806b-0ee6451a7021-kube-api-access-45lb8\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.432256 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.432285 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.432415 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.432660 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.432802 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.534868 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.534953 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.535028 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.535070 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.535107 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.535139 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-config-data\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.535262 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45lb8\" (UniqueName: \"kubernetes.io/projected/b0ec9924-8223-4a07-806b-0ee6451a7021-kube-api-access-45lb8\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.535301 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.535401 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.536129 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.536397 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.537178 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.537510 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-config-data\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.538018 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.547286 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.548232 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.548344 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.565338 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45lb8\" (UniqueName: \"kubernetes.io/projected/b0ec9924-8223-4a07-806b-0ee6451a7021-kube-api-access-45lb8\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.572432 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"tempest-tests-tempest\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " pod="openstack/tempest-tests-tempest" Dec 05 10:11:01 crc kubenswrapper[4815]: I1205 10:11:01.657200 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 05 10:11:02 crc kubenswrapper[4815]: I1205 10:11:02.121341 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 05 10:11:02 crc kubenswrapper[4815]: W1205 10:11:02.131553 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0ec9924_8223_4a07_806b_0ee6451a7021.slice/crio-c70aa725f20c3733c9256756deba886a63e7b8e6d47cea8c8d11ebd6211b2421 WatchSource:0}: Error finding container c70aa725f20c3733c9256756deba886a63e7b8e6d47cea8c8d11ebd6211b2421: Status 404 returned error can't find the container with id c70aa725f20c3733c9256756deba886a63e7b8e6d47cea8c8d11ebd6211b2421 Dec 05 10:11:02 crc kubenswrapper[4815]: I1205 10:11:02.138294 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:11:02 crc kubenswrapper[4815]: I1205 10:11:02.287576 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"b0ec9924-8223-4a07-806b-0ee6451a7021","Type":"ContainerStarted","Data":"c70aa725f20c3733c9256756deba886a63e7b8e6d47cea8c8d11ebd6211b2421"} Dec 05 10:11:09 crc kubenswrapper[4815]: I1205 10:11:09.420027 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:11:09 crc kubenswrapper[4815]: E1205 10:11:09.421372 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:11:24 crc kubenswrapper[4815]: I1205 10:11:24.419161 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:11:24 crc kubenswrapper[4815]: E1205 10:11:24.420182 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:11:35 crc kubenswrapper[4815]: I1205 10:11:35.419264 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:11:35 crc kubenswrapper[4815]: E1205 10:11:35.419999 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:11:47 crc kubenswrapper[4815]: E1205 10:11:47.839976 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 05 10:11:47 crc kubenswrapper[4815]: E1205 10:11:47.843564 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-45lb8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(b0ec9924-8223-4a07-806b-0ee6451a7021): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 10:11:47 crc kubenswrapper[4815]: E1205 10:11:47.844803 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="b0ec9924-8223-4a07-806b-0ee6451a7021" Dec 05 10:11:48 crc kubenswrapper[4815]: E1205 10:11:48.177037 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="b0ec9924-8223-4a07-806b-0ee6451a7021" Dec 05 10:11:48 crc kubenswrapper[4815]: I1205 10:11:48.419135 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:11:48 crc kubenswrapper[4815]: E1205 10:11:48.419717 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:12:01 crc kubenswrapper[4815]: I1205 10:12:01.431857 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:12:01 crc kubenswrapper[4815]: E1205 10:12:01.432632 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:12:04 crc kubenswrapper[4815]: I1205 10:12:04.130050 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 05 10:12:06 crc kubenswrapper[4815]: I1205 10:12:06.376923 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"b0ec9924-8223-4a07-806b-0ee6451a7021","Type":"ContainerStarted","Data":"2294af189294c56c3a263e2762c25a03ddca150493001b774200fbdeef9467e3"} Dec 05 10:12:06 crc kubenswrapper[4815]: I1205 10:12:06.403525 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.412698886 podStartE2EDuration="1m6.403476852s" podCreationTimestamp="2025-12-05 10:11:00 +0000 UTC" firstStartedPulling="2025-12-05 10:11:02.134078505 +0000 UTC m=+3861.012685342" lastFinishedPulling="2025-12-05 10:12:04.124856471 +0000 UTC m=+3923.003463308" observedRunningTime="2025-12-05 10:12:06.397327434 +0000 UTC m=+3925.275934271" watchObservedRunningTime="2025-12-05 10:12:06.403476852 +0000 UTC m=+3925.282083689" Dec 05 10:12:12 crc kubenswrapper[4815]: I1205 10:12:12.419521 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:12:12 crc kubenswrapper[4815]: E1205 10:12:12.420367 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:12:25 crc kubenswrapper[4815]: I1205 10:12:25.432253 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:12:25 crc kubenswrapper[4815]: E1205 10:12:25.433129 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:12:38 crc kubenswrapper[4815]: I1205 10:12:38.418913 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:12:38 crc kubenswrapper[4815]: E1205 10:12:38.419808 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:12:49 crc kubenswrapper[4815]: I1205 10:12:49.419395 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:12:49 crc kubenswrapper[4815]: E1205 10:12:49.420397 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:13:01 crc kubenswrapper[4815]: I1205 10:13:01.425703 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:13:01 crc kubenswrapper[4815]: E1205 10:13:01.426435 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:13:14 crc kubenswrapper[4815]: I1205 10:13:14.419195 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:13:14 crc kubenswrapper[4815]: E1205 10:13:14.419943 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:13:29 crc kubenswrapper[4815]: I1205 10:13:29.419634 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:13:30 crc kubenswrapper[4815]: I1205 10:13:30.166227 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"b302353b8ed6e7609eaaca87ebca2fffa6b7c967de913ce1b3341b9b586d2603"} Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.214823 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn"] Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.217695 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.221192 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.221871 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.235992 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn"] Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.251910 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0faa5a0-8f1d-4361-9b6e-19270274a526-secret-volume\") pod \"collect-profiles-29415495-mnjdn\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.251963 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0faa5a0-8f1d-4361-9b6e-19270274a526-config-volume\") pod \"collect-profiles-29415495-mnjdn\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.252013 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnngj\" (UniqueName: \"kubernetes.io/projected/c0faa5a0-8f1d-4361-9b6e-19270274a526-kube-api-access-xnngj\") pod \"collect-profiles-29415495-mnjdn\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.353602 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0faa5a0-8f1d-4361-9b6e-19270274a526-secret-volume\") pod \"collect-profiles-29415495-mnjdn\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.353645 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0faa5a0-8f1d-4361-9b6e-19270274a526-config-volume\") pod \"collect-profiles-29415495-mnjdn\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.353680 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnngj\" (UniqueName: \"kubernetes.io/projected/c0faa5a0-8f1d-4361-9b6e-19270274a526-kube-api-access-xnngj\") pod \"collect-profiles-29415495-mnjdn\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.354588 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0faa5a0-8f1d-4361-9b6e-19270274a526-config-volume\") pod \"collect-profiles-29415495-mnjdn\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.361553 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0faa5a0-8f1d-4361-9b6e-19270274a526-secret-volume\") pod \"collect-profiles-29415495-mnjdn\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.374277 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnngj\" (UniqueName: \"kubernetes.io/projected/c0faa5a0-8f1d-4361-9b6e-19270274a526-kube-api-access-xnngj\") pod \"collect-profiles-29415495-mnjdn\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:00 crc kubenswrapper[4815]: I1205 10:15:00.546601 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:01 crc kubenswrapper[4815]: I1205 10:15:01.963108 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn"] Dec 05 10:15:02 crc kubenswrapper[4815]: I1205 10:15:02.337707 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" event={"ID":"c0faa5a0-8f1d-4361-9b6e-19270274a526","Type":"ContainerStarted","Data":"6cd7b074384738e8198fe5a6d59d0910463f9ac4b5804ef4456b20668f6d1237"} Dec 05 10:15:03 crc kubenswrapper[4815]: I1205 10:15:03.347148 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" event={"ID":"c0faa5a0-8f1d-4361-9b6e-19270274a526","Type":"ContainerStarted","Data":"7cf8a5a9702fc36df6b57ba3866c3759b2350858158b676073e41928a90abf25"} Dec 05 10:15:03 crc kubenswrapper[4815]: I1205 10:15:03.371317 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" podStartSLOduration=3.371284898 podStartE2EDuration="3.371284898s" podCreationTimestamp="2025-12-05 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:15:03.365990644 +0000 UTC m=+4102.244597481" watchObservedRunningTime="2025-12-05 10:15:03.371284898 +0000 UTC m=+4102.249891735" Dec 05 10:15:04 crc kubenswrapper[4815]: I1205 10:15:04.356539 4815 generic.go:334] "Generic (PLEG): container finished" podID="c0faa5a0-8f1d-4361-9b6e-19270274a526" containerID="7cf8a5a9702fc36df6b57ba3866c3759b2350858158b676073e41928a90abf25" exitCode=0 Dec 05 10:15:04 crc kubenswrapper[4815]: I1205 10:15:04.356805 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" event={"ID":"c0faa5a0-8f1d-4361-9b6e-19270274a526","Type":"ContainerDied","Data":"7cf8a5a9702fc36df6b57ba3866c3759b2350858158b676073e41928a90abf25"} Dec 05 10:15:05 crc kubenswrapper[4815]: I1205 10:15:05.819147 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:05 crc kubenswrapper[4815]: I1205 10:15:05.824434 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0faa5a0-8f1d-4361-9b6e-19270274a526-secret-volume\") pod \"c0faa5a0-8f1d-4361-9b6e-19270274a526\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " Dec 05 10:15:05 crc kubenswrapper[4815]: I1205 10:15:05.824525 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0faa5a0-8f1d-4361-9b6e-19270274a526-config-volume\") pod \"c0faa5a0-8f1d-4361-9b6e-19270274a526\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " Dec 05 10:15:05 crc kubenswrapper[4815]: I1205 10:15:05.824598 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnngj\" (UniqueName: \"kubernetes.io/projected/c0faa5a0-8f1d-4361-9b6e-19270274a526-kube-api-access-xnngj\") pod \"c0faa5a0-8f1d-4361-9b6e-19270274a526\" (UID: \"c0faa5a0-8f1d-4361-9b6e-19270274a526\") " Dec 05 10:15:05 crc kubenswrapper[4815]: I1205 10:15:05.826120 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0faa5a0-8f1d-4361-9b6e-19270274a526-config-volume" (OuterVolumeSpecName: "config-volume") pod "c0faa5a0-8f1d-4361-9b6e-19270274a526" (UID: "c0faa5a0-8f1d-4361-9b6e-19270274a526"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:15:05 crc kubenswrapper[4815]: I1205 10:15:05.831944 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0faa5a0-8f1d-4361-9b6e-19270274a526-kube-api-access-xnngj" (OuterVolumeSpecName: "kube-api-access-xnngj") pod "c0faa5a0-8f1d-4361-9b6e-19270274a526" (UID: "c0faa5a0-8f1d-4361-9b6e-19270274a526"). InnerVolumeSpecName "kube-api-access-xnngj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:15:05 crc kubenswrapper[4815]: I1205 10:15:05.833688 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0faa5a0-8f1d-4361-9b6e-19270274a526-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c0faa5a0-8f1d-4361-9b6e-19270274a526" (UID: "c0faa5a0-8f1d-4361-9b6e-19270274a526"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:15:05 crc kubenswrapper[4815]: I1205 10:15:05.927533 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c0faa5a0-8f1d-4361-9b6e-19270274a526-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:15:05 crc kubenswrapper[4815]: I1205 10:15:05.927570 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c0faa5a0-8f1d-4361-9b6e-19270274a526-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:15:05 crc kubenswrapper[4815]: I1205 10:15:05.927582 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnngj\" (UniqueName: \"kubernetes.io/projected/c0faa5a0-8f1d-4361-9b6e-19270274a526-kube-api-access-xnngj\") on node \"crc\" DevicePath \"\"" Dec 05 10:15:06 crc kubenswrapper[4815]: I1205 10:15:06.377214 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" event={"ID":"c0faa5a0-8f1d-4361-9b6e-19270274a526","Type":"ContainerDied","Data":"6cd7b074384738e8198fe5a6d59d0910463f9ac4b5804ef4456b20668f6d1237"} Dec 05 10:15:06 crc kubenswrapper[4815]: I1205 10:15:06.377474 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cd7b074384738e8198fe5a6d59d0910463f9ac4b5804ef4456b20668f6d1237" Dec 05 10:15:06 crc kubenswrapper[4815]: I1205 10:15:06.377288 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415495-mnjdn" Dec 05 10:15:06 crc kubenswrapper[4815]: I1205 10:15:06.500544 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8"] Dec 05 10:15:06 crc kubenswrapper[4815]: I1205 10:15:06.512983 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415450-mxlg8"] Dec 05 10:15:07 crc kubenswrapper[4815]: I1205 10:15:07.435753 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b2189a3-9b89-4167-91ce-b9e92a360d11" path="/var/lib/kubelet/pods/0b2189a3-9b89-4167-91ce-b9e92a360d11/volumes" Dec 05 10:15:50 crc kubenswrapper[4815]: I1205 10:15:50.192243 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:15:50 crc kubenswrapper[4815]: I1205 10:15:50.192976 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:16:03 crc kubenswrapper[4815]: I1205 10:16:03.577448 4815 scope.go:117] "RemoveContainer" containerID="ff2675db2ab534dc3b503cf7eed5658aae5dbe001a549df516f0c67935bc53dd" Dec 05 10:16:20 crc kubenswrapper[4815]: I1205 10:16:20.192593 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:16:20 crc kubenswrapper[4815]: I1205 10:16:20.193043 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:16:50 crc kubenswrapper[4815]: I1205 10:16:50.192534 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:16:50 crc kubenswrapper[4815]: I1205 10:16:50.193114 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:16:50 crc kubenswrapper[4815]: I1205 10:16:50.193166 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 10:16:50 crc kubenswrapper[4815]: I1205 10:16:50.193894 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b302353b8ed6e7609eaaca87ebca2fffa6b7c967de913ce1b3341b9b586d2603"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:16:50 crc kubenswrapper[4815]: I1205 10:16:50.193946 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://b302353b8ed6e7609eaaca87ebca2fffa6b7c967de913ce1b3341b9b586d2603" gracePeriod=600 Dec 05 10:16:50 crc kubenswrapper[4815]: I1205 10:16:50.377588 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="b302353b8ed6e7609eaaca87ebca2fffa6b7c967de913ce1b3341b9b586d2603" exitCode=0 Dec 05 10:16:50 crc kubenswrapper[4815]: I1205 10:16:50.377632 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"b302353b8ed6e7609eaaca87ebca2fffa6b7c967de913ce1b3341b9b586d2603"} Dec 05 10:16:50 crc kubenswrapper[4815]: I1205 10:16:50.377682 4815 scope.go:117] "RemoveContainer" containerID="9e36831947dd3dba2e52896152eef41dc970bc8b0d11fbd9231f396b94bb58d1" Dec 05 10:16:51 crc kubenswrapper[4815]: I1205 10:16:51.389387 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739"} Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.699796 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-m26ds"] Dec 05 10:17:11 crc kubenswrapper[4815]: E1205 10:17:11.700964 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0faa5a0-8f1d-4361-9b6e-19270274a526" containerName="collect-profiles" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.701000 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0faa5a0-8f1d-4361-9b6e-19270274a526" containerName="collect-profiles" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.701281 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0faa5a0-8f1d-4361-9b6e-19270274a526" containerName="collect-profiles" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.703476 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.744675 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m26ds"] Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.789106 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-catalog-content\") pod \"certified-operators-m26ds\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.789162 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phbgd\" (UniqueName: \"kubernetes.io/projected/d3197424-c356-40cd-b7c7-bda870708a20-kube-api-access-phbgd\") pod \"certified-operators-m26ds\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.789195 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-utilities\") pod \"certified-operators-m26ds\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.890133 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-catalog-content\") pod \"certified-operators-m26ds\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.890206 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phbgd\" (UniqueName: \"kubernetes.io/projected/d3197424-c356-40cd-b7c7-bda870708a20-kube-api-access-phbgd\") pod \"certified-operators-m26ds\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.890238 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-utilities\") pod \"certified-operators-m26ds\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.890712 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-catalog-content\") pod \"certified-operators-m26ds\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.890763 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-utilities\") pod \"certified-operators-m26ds\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:11 crc kubenswrapper[4815]: I1205 10:17:11.909096 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phbgd\" (UniqueName: \"kubernetes.io/projected/d3197424-c356-40cd-b7c7-bda870708a20-kube-api-access-phbgd\") pod \"certified-operators-m26ds\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:12 crc kubenswrapper[4815]: I1205 10:17:12.025165 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:12 crc kubenswrapper[4815]: I1205 10:17:12.662618 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-m26ds"] Dec 05 10:17:13 crc kubenswrapper[4815]: I1205 10:17:13.595230 4815 generic.go:334] "Generic (PLEG): container finished" podID="d3197424-c356-40cd-b7c7-bda870708a20" containerID="3e0b02f04394652c2eef46171813700c0cdfd9557ea12d940f5344d1cf6ffdd3" exitCode=0 Dec 05 10:17:13 crc kubenswrapper[4815]: I1205 10:17:13.595591 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m26ds" event={"ID":"d3197424-c356-40cd-b7c7-bda870708a20","Type":"ContainerDied","Data":"3e0b02f04394652c2eef46171813700c0cdfd9557ea12d940f5344d1cf6ffdd3"} Dec 05 10:17:13 crc kubenswrapper[4815]: I1205 10:17:13.595624 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m26ds" event={"ID":"d3197424-c356-40cd-b7c7-bda870708a20","Type":"ContainerStarted","Data":"6ae96888b6169e05d2ffcd6508db0d3ad21650ac1559e315869b204b1b2e8fcd"} Dec 05 10:17:13 crc kubenswrapper[4815]: I1205 10:17:13.597955 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:17:14 crc kubenswrapper[4815]: I1205 10:17:14.612763 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m26ds" event={"ID":"d3197424-c356-40cd-b7c7-bda870708a20","Type":"ContainerStarted","Data":"501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19"} Dec 05 10:17:16 crc kubenswrapper[4815]: E1205 10:17:16.875078 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3197424_c356_40cd_b7c7_bda870708a20.slice/crio-501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3197424_c356_40cd_b7c7_bda870708a20.slice/crio-conmon-501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19.scope\": RecentStats: unable to find data in memory cache]" Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.561288 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q4blr"] Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.564083 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.591069 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4blr"] Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.651945 4815 generic.go:334] "Generic (PLEG): container finished" podID="d3197424-c356-40cd-b7c7-bda870708a20" containerID="501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19" exitCode=0 Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.652041 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m26ds" event={"ID":"d3197424-c356-40cd-b7c7-bda870708a20","Type":"ContainerDied","Data":"501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19"} Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.749957 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbjvv\" (UniqueName: \"kubernetes.io/projected/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-kube-api-access-gbjvv\") pod \"redhat-marketplace-q4blr\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.750006 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-utilities\") pod \"redhat-marketplace-q4blr\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.750123 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-catalog-content\") pod \"redhat-marketplace-q4blr\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.852188 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-catalog-content\") pod \"redhat-marketplace-q4blr\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.852372 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbjvv\" (UniqueName: \"kubernetes.io/projected/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-kube-api-access-gbjvv\") pod \"redhat-marketplace-q4blr\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.852418 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-utilities\") pod \"redhat-marketplace-q4blr\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.852786 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-catalog-content\") pod \"redhat-marketplace-q4blr\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.852840 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-utilities\") pod \"redhat-marketplace-q4blr\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.879476 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbjvv\" (UniqueName: \"kubernetes.io/projected/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-kube-api-access-gbjvv\") pod \"redhat-marketplace-q4blr\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:17 crc kubenswrapper[4815]: I1205 10:17:17.889200 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:18 crc kubenswrapper[4815]: I1205 10:17:18.508033 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4blr"] Dec 05 10:17:18 crc kubenswrapper[4815]: W1205 10:17:18.522062 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod436ffd5a_d858_43a5_8d40_456b5c8dbf3c.slice/crio-3bd4aff0260734076582272d7098339e6f9dadbcfb5ba917bb92bbcc4bbadc07 WatchSource:0}: Error finding container 3bd4aff0260734076582272d7098339e6f9dadbcfb5ba917bb92bbcc4bbadc07: Status 404 returned error can't find the container with id 3bd4aff0260734076582272d7098339e6f9dadbcfb5ba917bb92bbcc4bbadc07 Dec 05 10:17:18 crc kubenswrapper[4815]: I1205 10:17:18.669962 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m26ds" event={"ID":"d3197424-c356-40cd-b7c7-bda870708a20","Type":"ContainerStarted","Data":"33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f"} Dec 05 10:17:18 crc kubenswrapper[4815]: I1205 10:17:18.673868 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4blr" event={"ID":"436ffd5a-d858-43a5-8d40-456b5c8dbf3c","Type":"ContainerStarted","Data":"3bd4aff0260734076582272d7098339e6f9dadbcfb5ba917bb92bbcc4bbadc07"} Dec 05 10:17:18 crc kubenswrapper[4815]: I1205 10:17:18.705556 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-m26ds" podStartSLOduration=3.2623680999999998 podStartE2EDuration="7.705523934s" podCreationTimestamp="2025-12-05 10:17:11 +0000 UTC" firstStartedPulling="2025-12-05 10:17:13.597532741 +0000 UTC m=+4232.476139578" lastFinishedPulling="2025-12-05 10:17:18.040688575 +0000 UTC m=+4236.919295412" observedRunningTime="2025-12-05 10:17:18.693146246 +0000 UTC m=+4237.571753083" watchObservedRunningTime="2025-12-05 10:17:18.705523934 +0000 UTC m=+4237.584130771" Dec 05 10:17:19 crc kubenswrapper[4815]: I1205 10:17:19.727012 4815 generic.go:334] "Generic (PLEG): container finished" podID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" containerID="b9069174ab4ee2c3dd22a2eb5ec67d55c494cabb1eac33710a56ebe8d5794ad4" exitCode=0 Dec 05 10:17:19 crc kubenswrapper[4815]: I1205 10:17:19.727300 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4blr" event={"ID":"436ffd5a-d858-43a5-8d40-456b5c8dbf3c","Type":"ContainerDied","Data":"b9069174ab4ee2c3dd22a2eb5ec67d55c494cabb1eac33710a56ebe8d5794ad4"} Dec 05 10:17:20 crc kubenswrapper[4815]: I1205 10:17:20.737214 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4blr" event={"ID":"436ffd5a-d858-43a5-8d40-456b5c8dbf3c","Type":"ContainerStarted","Data":"5c35c02d722f07073d50f2d79e4bbbd9e11f90aebc5f12e181d989e5146481d7"} Dec 05 10:17:22 crc kubenswrapper[4815]: I1205 10:17:22.026184 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:22 crc kubenswrapper[4815]: I1205 10:17:22.027442 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:22 crc kubenswrapper[4815]: I1205 10:17:22.755575 4815 generic.go:334] "Generic (PLEG): container finished" podID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" containerID="5c35c02d722f07073d50f2d79e4bbbd9e11f90aebc5f12e181d989e5146481d7" exitCode=0 Dec 05 10:17:22 crc kubenswrapper[4815]: I1205 10:17:22.755678 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4blr" event={"ID":"436ffd5a-d858-43a5-8d40-456b5c8dbf3c","Type":"ContainerDied","Data":"5c35c02d722f07073d50f2d79e4bbbd9e11f90aebc5f12e181d989e5146481d7"} Dec 05 10:17:23 crc kubenswrapper[4815]: I1205 10:17:23.077756 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-m26ds" podUID="d3197424-c356-40cd-b7c7-bda870708a20" containerName="registry-server" probeResult="failure" output=< Dec 05 10:17:23 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 10:17:23 crc kubenswrapper[4815]: > Dec 05 10:17:23 crc kubenswrapper[4815]: I1205 10:17:23.776446 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4blr" event={"ID":"436ffd5a-d858-43a5-8d40-456b5c8dbf3c","Type":"ContainerStarted","Data":"dcabbc169b167f245c0927004045eb8170d4faff1d9a6ca363f67a26ebddd89d"} Dec 05 10:17:23 crc kubenswrapper[4815]: I1205 10:17:23.908220 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q4blr" podStartSLOduration=3.488816739 podStartE2EDuration="6.908197595s" podCreationTimestamp="2025-12-05 10:17:17 +0000 UTC" firstStartedPulling="2025-12-05 10:17:19.737473066 +0000 UTC m=+4238.616079903" lastFinishedPulling="2025-12-05 10:17:23.156853922 +0000 UTC m=+4242.035460759" observedRunningTime="2025-12-05 10:17:23.801107637 +0000 UTC m=+4242.679714504" watchObservedRunningTime="2025-12-05 10:17:23.908197595 +0000 UTC m=+4242.786804432" Dec 05 10:17:27 crc kubenswrapper[4815]: I1205 10:17:27.889869 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:27 crc kubenswrapper[4815]: I1205 10:17:27.891453 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:27 crc kubenswrapper[4815]: I1205 10:17:27.958939 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:28 crc kubenswrapper[4815]: I1205 10:17:28.678882 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:28 crc kubenswrapper[4815]: I1205 10:17:28.727622 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4blr"] Dec 05 10:17:30 crc kubenswrapper[4815]: I1205 10:17:30.631752 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q4blr" podUID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" containerName="registry-server" containerID="cri-o://dcabbc169b167f245c0927004045eb8170d4faff1d9a6ca363f67a26ebddd89d" gracePeriod=2 Dec 05 10:17:31 crc kubenswrapper[4815]: I1205 10:17:31.645987 4815 generic.go:334] "Generic (PLEG): container finished" podID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" containerID="dcabbc169b167f245c0927004045eb8170d4faff1d9a6ca363f67a26ebddd89d" exitCode=0 Dec 05 10:17:31 crc kubenswrapper[4815]: I1205 10:17:31.646186 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4blr" event={"ID":"436ffd5a-d858-43a5-8d40-456b5c8dbf3c","Type":"ContainerDied","Data":"dcabbc169b167f245c0927004045eb8170d4faff1d9a6ca363f67a26ebddd89d"} Dec 05 10:17:31 crc kubenswrapper[4815]: I1205 10:17:31.876080 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:31 crc kubenswrapper[4815]: I1205 10:17:31.997538 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-utilities\") pod \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " Dec 05 10:17:31 crc kubenswrapper[4815]: I1205 10:17:31.997633 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-catalog-content\") pod \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " Dec 05 10:17:31 crc kubenswrapper[4815]: I1205 10:17:31.997842 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbjvv\" (UniqueName: \"kubernetes.io/projected/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-kube-api-access-gbjvv\") pod \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\" (UID: \"436ffd5a-d858-43a5-8d40-456b5c8dbf3c\") " Dec 05 10:17:31 crc kubenswrapper[4815]: I1205 10:17:31.998274 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-utilities" (OuterVolumeSpecName: "utilities") pod "436ffd5a-d858-43a5-8d40-456b5c8dbf3c" (UID: "436ffd5a-d858-43a5-8d40-456b5c8dbf3c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:17:31 crc kubenswrapper[4815]: I1205 10:17:31.998387 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.005629 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-kube-api-access-gbjvv" (OuterVolumeSpecName: "kube-api-access-gbjvv") pod "436ffd5a-d858-43a5-8d40-456b5c8dbf3c" (UID: "436ffd5a-d858-43a5-8d40-456b5c8dbf3c"). InnerVolumeSpecName "kube-api-access-gbjvv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.017389 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "436ffd5a-d858-43a5-8d40-456b5c8dbf3c" (UID: "436ffd5a-d858-43a5-8d40-456b5c8dbf3c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.080337 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.100863 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.100892 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbjvv\" (UniqueName: \"kubernetes.io/projected/436ffd5a-d858-43a5-8d40-456b5c8dbf3c-kube-api-access-gbjvv\") on node \"crc\" DevicePath \"\"" Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.132832 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.658062 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q4blr" event={"ID":"436ffd5a-d858-43a5-8d40-456b5c8dbf3c","Type":"ContainerDied","Data":"3bd4aff0260734076582272d7098339e6f9dadbcfb5ba917bb92bbcc4bbadc07"} Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.658149 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q4blr" Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.658163 4815 scope.go:117] "RemoveContainer" containerID="dcabbc169b167f245c0927004045eb8170d4faff1d9a6ca363f67a26ebddd89d" Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.688463 4815 scope.go:117] "RemoveContainer" containerID="5c35c02d722f07073d50f2d79e4bbbd9e11f90aebc5f12e181d989e5146481d7" Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.700759 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4blr"] Dec 05 10:17:32 crc kubenswrapper[4815]: I1205 10:17:32.711615 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q4blr"] Dec 05 10:17:33 crc kubenswrapper[4815]: I1205 10:17:33.141131 4815 scope.go:117] "RemoveContainer" containerID="b9069174ab4ee2c3dd22a2eb5ec67d55c494cabb1eac33710a56ebe8d5794ad4" Dec 05 10:17:33 crc kubenswrapper[4815]: I1205 10:17:33.436445 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" path="/var/lib/kubelet/pods/436ffd5a-d858-43a5-8d40-456b5c8dbf3c/volumes" Dec 05 10:17:33 crc kubenswrapper[4815]: I1205 10:17:33.517207 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m26ds"] Dec 05 10:17:33 crc kubenswrapper[4815]: I1205 10:17:33.670643 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-m26ds" podUID="d3197424-c356-40cd-b7c7-bda870708a20" containerName="registry-server" containerID="cri-o://33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f" gracePeriod=2 Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.232770 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.355293 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-catalog-content\") pod \"d3197424-c356-40cd-b7c7-bda870708a20\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.355356 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phbgd\" (UniqueName: \"kubernetes.io/projected/d3197424-c356-40cd-b7c7-bda870708a20-kube-api-access-phbgd\") pod \"d3197424-c356-40cd-b7c7-bda870708a20\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.355396 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-utilities\") pod \"d3197424-c356-40cd-b7c7-bda870708a20\" (UID: \"d3197424-c356-40cd-b7c7-bda870708a20\") " Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.356265 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-utilities" (OuterVolumeSpecName: "utilities") pod "d3197424-c356-40cd-b7c7-bda870708a20" (UID: "d3197424-c356-40cd-b7c7-bda870708a20"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.357292 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.362512 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3197424-c356-40cd-b7c7-bda870708a20-kube-api-access-phbgd" (OuterVolumeSpecName: "kube-api-access-phbgd") pod "d3197424-c356-40cd-b7c7-bda870708a20" (UID: "d3197424-c356-40cd-b7c7-bda870708a20"). InnerVolumeSpecName "kube-api-access-phbgd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.394691 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d3197424-c356-40cd-b7c7-bda870708a20" (UID: "d3197424-c356-40cd-b7c7-bda870708a20"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.459592 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d3197424-c356-40cd-b7c7-bda870708a20-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.459636 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phbgd\" (UniqueName: \"kubernetes.io/projected/d3197424-c356-40cd-b7c7-bda870708a20-kube-api-access-phbgd\") on node \"crc\" DevicePath \"\"" Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.681239 4815 generic.go:334] "Generic (PLEG): container finished" podID="d3197424-c356-40cd-b7c7-bda870708a20" containerID="33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f" exitCode=0 Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.681287 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m26ds" event={"ID":"d3197424-c356-40cd-b7c7-bda870708a20","Type":"ContainerDied","Data":"33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f"} Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.681319 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-m26ds" event={"ID":"d3197424-c356-40cd-b7c7-bda870708a20","Type":"ContainerDied","Data":"6ae96888b6169e05d2ffcd6508db0d3ad21650ac1559e315869b204b1b2e8fcd"} Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.681342 4815 scope.go:117] "RemoveContainer" containerID="33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f" Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.681344 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-m26ds" Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.723404 4815 scope.go:117] "RemoveContainer" containerID="501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19" Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.726733 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-m26ds"] Dec 05 10:17:34 crc kubenswrapper[4815]: I1205 10:17:34.738021 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-m26ds"] Dec 05 10:17:35 crc kubenswrapper[4815]: I1205 10:17:35.042915 4815 scope.go:117] "RemoveContainer" containerID="3e0b02f04394652c2eef46171813700c0cdfd9557ea12d940f5344d1cf6ffdd3" Dec 05 10:17:35 crc kubenswrapper[4815]: I1205 10:17:35.180913 4815 scope.go:117] "RemoveContainer" containerID="33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f" Dec 05 10:17:35 crc kubenswrapper[4815]: E1205 10:17:35.182664 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f\": container with ID starting with 33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f not found: ID does not exist" containerID="33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f" Dec 05 10:17:35 crc kubenswrapper[4815]: I1205 10:17:35.182731 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f"} err="failed to get container status \"33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f\": rpc error: code = NotFound desc = could not find container \"33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f\": container with ID starting with 33a2ac21b6b58265544d4e48639314a64c70dbd9999d682df51299b0bd05223f not found: ID does not exist" Dec 05 10:17:35 crc kubenswrapper[4815]: I1205 10:17:35.182763 4815 scope.go:117] "RemoveContainer" containerID="501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19" Dec 05 10:17:35 crc kubenswrapper[4815]: E1205 10:17:35.183388 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19\": container with ID starting with 501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19 not found: ID does not exist" containerID="501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19" Dec 05 10:17:35 crc kubenswrapper[4815]: I1205 10:17:35.183416 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19"} err="failed to get container status \"501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19\": rpc error: code = NotFound desc = could not find container \"501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19\": container with ID starting with 501e6e0573667b2d2132c33f5efc28195f8c171361fb57aa9990ba2a167c6c19 not found: ID does not exist" Dec 05 10:17:35 crc kubenswrapper[4815]: I1205 10:17:35.183437 4815 scope.go:117] "RemoveContainer" containerID="3e0b02f04394652c2eef46171813700c0cdfd9557ea12d940f5344d1cf6ffdd3" Dec 05 10:17:35 crc kubenswrapper[4815]: E1205 10:17:35.183885 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e0b02f04394652c2eef46171813700c0cdfd9557ea12d940f5344d1cf6ffdd3\": container with ID starting with 3e0b02f04394652c2eef46171813700c0cdfd9557ea12d940f5344d1cf6ffdd3 not found: ID does not exist" containerID="3e0b02f04394652c2eef46171813700c0cdfd9557ea12d940f5344d1cf6ffdd3" Dec 05 10:17:35 crc kubenswrapper[4815]: I1205 10:17:35.183917 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e0b02f04394652c2eef46171813700c0cdfd9557ea12d940f5344d1cf6ffdd3"} err="failed to get container status \"3e0b02f04394652c2eef46171813700c0cdfd9557ea12d940f5344d1cf6ffdd3\": rpc error: code = NotFound desc = could not find container \"3e0b02f04394652c2eef46171813700c0cdfd9557ea12d940f5344d1cf6ffdd3\": container with ID starting with 3e0b02f04394652c2eef46171813700c0cdfd9557ea12d940f5344d1cf6ffdd3 not found: ID does not exist" Dec 05 10:17:35 crc kubenswrapper[4815]: I1205 10:17:35.431234 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3197424-c356-40cd-b7c7-bda870708a20" path="/var/lib/kubelet/pods/d3197424-c356-40cd-b7c7-bda870708a20/volumes" Dec 05 10:18:04 crc kubenswrapper[4815]: I1205 10:18:04.854709 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" podUID="46274b62-2000-4c62-8688-af6b5fcb87dc" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.49:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 10:18:05 crc kubenswrapper[4815]: I1205 10:18:05.226702 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" podUID="4abce90f-180f-4302-aecd-89bf5739757c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 10:18:18 crc kubenswrapper[4815]: I1205 10:18:18.070015 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-1012-account-create-update-gd8rc"] Dec 05 10:18:18 crc kubenswrapper[4815]: I1205 10:18:18.078953 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-s8qsm"] Dec 05 10:18:18 crc kubenswrapper[4815]: I1205 10:18:18.087380 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-1012-account-create-update-gd8rc"] Dec 05 10:18:18 crc kubenswrapper[4815]: I1205 10:18:18.095286 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-s8qsm"] Dec 05 10:18:19 crc kubenswrapper[4815]: I1205 10:18:19.469527 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7170277e-ef0a-4821-9c74-0c61204feabb" path="/var/lib/kubelet/pods/7170277e-ef0a-4821-9c74-0c61204feabb/volumes" Dec 05 10:18:19 crc kubenswrapper[4815]: I1205 10:18:19.470590 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdb34d69-b26b-4258-9719-50a69ddce407" path="/var/lib/kubelet/pods/fdb34d69-b26b-4258-9719-50a69ddce407/volumes" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.415727 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-snnnj"] Dec 05 10:18:34 crc kubenswrapper[4815]: E1205 10:18:34.416588 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" containerName="registry-server" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.416604 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" containerName="registry-server" Dec 05 10:18:34 crc kubenswrapper[4815]: E1205 10:18:34.416652 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3197424-c356-40cd-b7c7-bda870708a20" containerName="registry-server" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.416663 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3197424-c356-40cd-b7c7-bda870708a20" containerName="registry-server" Dec 05 10:18:34 crc kubenswrapper[4815]: E1205 10:18:34.416681 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" containerName="extract-content" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.416687 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" containerName="extract-content" Dec 05 10:18:34 crc kubenswrapper[4815]: E1205 10:18:34.416702 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" containerName="extract-utilities" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.416709 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" containerName="extract-utilities" Dec 05 10:18:34 crc kubenswrapper[4815]: E1205 10:18:34.416718 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3197424-c356-40cd-b7c7-bda870708a20" containerName="extract-utilities" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.416725 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3197424-c356-40cd-b7c7-bda870708a20" containerName="extract-utilities" Dec 05 10:18:34 crc kubenswrapper[4815]: E1205 10:18:34.416743 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3197424-c356-40cd-b7c7-bda870708a20" containerName="extract-content" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.416755 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3197424-c356-40cd-b7c7-bda870708a20" containerName="extract-content" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.417019 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3197424-c356-40cd-b7c7-bda870708a20" containerName="registry-server" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.417043 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="436ffd5a-d858-43a5-8d40-456b5c8dbf3c" containerName="registry-server" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.418806 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.435960 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-snnnj"] Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.491813 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-catalog-content\") pod \"redhat-operators-snnnj\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.492051 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fslw\" (UniqueName: \"kubernetes.io/projected/ce6d12fe-7de8-459f-95e4-b103bb1febb3-kube-api-access-9fslw\") pod \"redhat-operators-snnnj\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.492129 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-utilities\") pod \"redhat-operators-snnnj\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.593674 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-utilities\") pod \"redhat-operators-snnnj\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.593803 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-catalog-content\") pod \"redhat-operators-snnnj\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.593909 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fslw\" (UniqueName: \"kubernetes.io/projected/ce6d12fe-7de8-459f-95e4-b103bb1febb3-kube-api-access-9fslw\") pod \"redhat-operators-snnnj\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.594169 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-utilities\") pod \"redhat-operators-snnnj\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.594424 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-catalog-content\") pod \"redhat-operators-snnnj\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.636892 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fslw\" (UniqueName: \"kubernetes.io/projected/ce6d12fe-7de8-459f-95e4-b103bb1febb3-kube-api-access-9fslw\") pod \"redhat-operators-snnnj\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:34 crc kubenswrapper[4815]: I1205 10:18:34.771222 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:35 crc kubenswrapper[4815]: I1205 10:18:35.296975 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-snnnj"] Dec 05 10:18:36 crc kubenswrapper[4815]: I1205 10:18:36.288401 4815 generic.go:334] "Generic (PLEG): container finished" podID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerID="d9241c0b4abcd7c6cfe327c0dea411a743212842d76a7d11b84fc602f59fb5e5" exitCode=0 Dec 05 10:18:36 crc kubenswrapper[4815]: I1205 10:18:36.289069 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snnnj" event={"ID":"ce6d12fe-7de8-459f-95e4-b103bb1febb3","Type":"ContainerDied","Data":"d9241c0b4abcd7c6cfe327c0dea411a743212842d76a7d11b84fc602f59fb5e5"} Dec 05 10:18:36 crc kubenswrapper[4815]: I1205 10:18:36.289107 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snnnj" event={"ID":"ce6d12fe-7de8-459f-95e4-b103bb1febb3","Type":"ContainerStarted","Data":"e43618ee88c9bb87814213a3edda719ebb53ebb83c76dae46e7fe63e1deb8e91"} Dec 05 10:18:38 crc kubenswrapper[4815]: I1205 10:18:38.331927 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snnnj" event={"ID":"ce6d12fe-7de8-459f-95e4-b103bb1febb3","Type":"ContainerStarted","Data":"34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4"} Dec 05 10:18:42 crc kubenswrapper[4815]: I1205 10:18:42.366476 4815 generic.go:334] "Generic (PLEG): container finished" podID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerID="34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4" exitCode=0 Dec 05 10:18:42 crc kubenswrapper[4815]: I1205 10:18:42.366531 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snnnj" event={"ID":"ce6d12fe-7de8-459f-95e4-b103bb1febb3","Type":"ContainerDied","Data":"34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4"} Dec 05 10:18:44 crc kubenswrapper[4815]: I1205 10:18:44.392026 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snnnj" event={"ID":"ce6d12fe-7de8-459f-95e4-b103bb1febb3","Type":"ContainerStarted","Data":"92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db"} Dec 05 10:18:44 crc kubenswrapper[4815]: I1205 10:18:44.430221 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-snnnj" podStartSLOduration=3.569026341 podStartE2EDuration="10.430207306s" podCreationTimestamp="2025-12-05 10:18:34 +0000 UTC" firstStartedPulling="2025-12-05 10:18:36.291210541 +0000 UTC m=+4315.169817378" lastFinishedPulling="2025-12-05 10:18:43.152391516 +0000 UTC m=+4322.030998343" observedRunningTime="2025-12-05 10:18:44.413053207 +0000 UTC m=+4323.291660044" watchObservedRunningTime="2025-12-05 10:18:44.430207306 +0000 UTC m=+4323.308814143" Dec 05 10:18:44 crc kubenswrapper[4815]: I1205 10:18:44.771546 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:44 crc kubenswrapper[4815]: I1205 10:18:44.772673 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:45 crc kubenswrapper[4815]: I1205 10:18:45.823714 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-snnnj" podUID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerName="registry-server" probeResult="failure" output=< Dec 05 10:18:45 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 10:18:45 crc kubenswrapper[4815]: > Dec 05 10:18:50 crc kubenswrapper[4815]: I1205 10:18:50.192913 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:18:50 crc kubenswrapper[4815]: I1205 10:18:50.194241 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:18:54 crc kubenswrapper[4815]: I1205 10:18:54.823447 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:54 crc kubenswrapper[4815]: I1205 10:18:54.887362 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:55 crc kubenswrapper[4815]: I1205 10:18:55.068220 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-snnnj"] Dec 05 10:18:56 crc kubenswrapper[4815]: I1205 10:18:56.519647 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-snnnj" podUID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerName="registry-server" containerID="cri-o://92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db" gracePeriod=2 Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.099312 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.241388 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-catalog-content\") pod \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.241752 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-utilities\") pod \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.241868 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fslw\" (UniqueName: \"kubernetes.io/projected/ce6d12fe-7de8-459f-95e4-b103bb1febb3-kube-api-access-9fslw\") pod \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\" (UID: \"ce6d12fe-7de8-459f-95e4-b103bb1febb3\") " Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.242395 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-utilities" (OuterVolumeSpecName: "utilities") pod "ce6d12fe-7de8-459f-95e4-b103bb1febb3" (UID: "ce6d12fe-7de8-459f-95e4-b103bb1febb3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.343759 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.352159 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce6d12fe-7de8-459f-95e4-b103bb1febb3" (UID: "ce6d12fe-7de8-459f-95e4-b103bb1febb3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.445687 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce6d12fe-7de8-459f-95e4-b103bb1febb3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.531056 4815 generic.go:334] "Generic (PLEG): container finished" podID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerID="92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db" exitCode=0 Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.531154 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snnnj" event={"ID":"ce6d12fe-7de8-459f-95e4-b103bb1febb3","Type":"ContainerDied","Data":"92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db"} Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.531198 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-snnnj" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.531225 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-snnnj" event={"ID":"ce6d12fe-7de8-459f-95e4-b103bb1febb3","Type":"ContainerDied","Data":"e43618ee88c9bb87814213a3edda719ebb53ebb83c76dae46e7fe63e1deb8e91"} Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.531297 4815 scope.go:117] "RemoveContainer" containerID="92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.561283 4815 scope.go:117] "RemoveContainer" containerID="34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.726316 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce6d12fe-7de8-459f-95e4-b103bb1febb3-kube-api-access-9fslw" (OuterVolumeSpecName: "kube-api-access-9fslw") pod "ce6d12fe-7de8-459f-95e4-b103bb1febb3" (UID: "ce6d12fe-7de8-459f-95e4-b103bb1febb3"). InnerVolumeSpecName "kube-api-access-9fslw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.752124 4815 scope.go:117] "RemoveContainer" containerID="d9241c0b4abcd7c6cfe327c0dea411a743212842d76a7d11b84fc602f59fb5e5" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.755324 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9fslw\" (UniqueName: \"kubernetes.io/projected/ce6d12fe-7de8-459f-95e4-b103bb1febb3-kube-api-access-9fslw\") on node \"crc\" DevicePath \"\"" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.844934 4815 scope.go:117] "RemoveContainer" containerID="92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db" Dec 05 10:18:57 crc kubenswrapper[4815]: E1205 10:18:57.845409 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db\": container with ID starting with 92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db not found: ID does not exist" containerID="92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.845446 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db"} err="failed to get container status \"92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db\": rpc error: code = NotFound desc = could not find container \"92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db\": container with ID starting with 92a448d220d693dc2b7b33842ad2aabb91e7939685b2d9a196ef25f4ad3eb6db not found: ID does not exist" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.845475 4815 scope.go:117] "RemoveContainer" containerID="34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4" Dec 05 10:18:57 crc kubenswrapper[4815]: E1205 10:18:57.845727 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4\": container with ID starting with 34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4 not found: ID does not exist" containerID="34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.845755 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4"} err="failed to get container status \"34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4\": rpc error: code = NotFound desc = could not find container \"34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4\": container with ID starting with 34c6769ea96af3df8564ccf7b10ac88d34123e04b15e2460e16a835a174c5cb4 not found: ID does not exist" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.845772 4815 scope.go:117] "RemoveContainer" containerID="d9241c0b4abcd7c6cfe327c0dea411a743212842d76a7d11b84fc602f59fb5e5" Dec 05 10:18:57 crc kubenswrapper[4815]: E1205 10:18:57.846019 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9241c0b4abcd7c6cfe327c0dea411a743212842d76a7d11b84fc602f59fb5e5\": container with ID starting with d9241c0b4abcd7c6cfe327c0dea411a743212842d76a7d11b84fc602f59fb5e5 not found: ID does not exist" containerID="d9241c0b4abcd7c6cfe327c0dea411a743212842d76a7d11b84fc602f59fb5e5" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.846048 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9241c0b4abcd7c6cfe327c0dea411a743212842d76a7d11b84fc602f59fb5e5"} err="failed to get container status \"d9241c0b4abcd7c6cfe327c0dea411a743212842d76a7d11b84fc602f59fb5e5\": rpc error: code = NotFound desc = could not find container \"d9241c0b4abcd7c6cfe327c0dea411a743212842d76a7d11b84fc602f59fb5e5\": container with ID starting with d9241c0b4abcd7c6cfe327c0dea411a743212842d76a7d11b84fc602f59fb5e5 not found: ID does not exist" Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.907289 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-snnnj"] Dec 05 10:18:57 crc kubenswrapper[4815]: I1205 10:18:57.917847 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-snnnj"] Dec 05 10:18:58 crc kubenswrapper[4815]: I1205 10:18:58.042081 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-l5xf8"] Dec 05 10:18:58 crc kubenswrapper[4815]: I1205 10:18:58.049729 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-l5xf8"] Dec 05 10:18:59 crc kubenswrapper[4815]: I1205 10:18:59.439420 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="733c76db-715e-41b2-a91f-4032c25e2ab9" path="/var/lib/kubelet/pods/733c76db-715e-41b2-a91f-4032c25e2ab9/volumes" Dec 05 10:18:59 crc kubenswrapper[4815]: I1205 10:18:59.445209 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" path="/var/lib/kubelet/pods/ce6d12fe-7de8-459f-95e4-b103bb1febb3/volumes" Dec 05 10:19:03 crc kubenswrapper[4815]: I1205 10:19:03.731616 4815 scope.go:117] "RemoveContainer" containerID="eebf0bf4b011d6ff3e9f4e7bcbe54282186f36f46fdebc15578a012747b13ee3" Dec 05 10:19:03 crc kubenswrapper[4815]: I1205 10:19:03.786030 4815 scope.go:117] "RemoveContainer" containerID="f7b9227bdcbb51659cdcb9e31207a69936a648a3c9df4cb86dc8f35e403e7627" Dec 05 10:19:03 crc kubenswrapper[4815]: I1205 10:19:03.852170 4815 scope.go:117] "RemoveContainer" containerID="57fc7ec8e61b912927d3065c9dce1ced121427a9daa73cb6b5e38572849004fc" Dec 05 10:19:20 crc kubenswrapper[4815]: I1205 10:19:20.192587 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:19:20 crc kubenswrapper[4815]: I1205 10:19:20.193106 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:19:42 crc kubenswrapper[4815]: I1205 10:19:42.968201 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-2fbnj"] Dec 05 10:19:42 crc kubenswrapper[4815]: E1205 10:19:42.969296 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerName="registry-server" Dec 05 10:19:42 crc kubenswrapper[4815]: I1205 10:19:42.969327 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerName="registry-server" Dec 05 10:19:42 crc kubenswrapper[4815]: E1205 10:19:42.969363 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerName="extract-utilities" Dec 05 10:19:42 crc kubenswrapper[4815]: I1205 10:19:42.969372 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerName="extract-utilities" Dec 05 10:19:42 crc kubenswrapper[4815]: E1205 10:19:42.969390 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerName="extract-content" Dec 05 10:19:42 crc kubenswrapper[4815]: I1205 10:19:42.969397 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerName="extract-content" Dec 05 10:19:42 crc kubenswrapper[4815]: I1205 10:19:42.969712 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce6d12fe-7de8-459f-95e4-b103bb1febb3" containerName="registry-server" Dec 05 10:19:42 crc kubenswrapper[4815]: I1205 10:19:42.991427 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2fbnj"] Dec 05 10:19:42 crc kubenswrapper[4815]: I1205 10:19:42.992425 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:43 crc kubenswrapper[4815]: I1205 10:19:43.221190 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-utilities\") pod \"community-operators-2fbnj\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:43 crc kubenswrapper[4815]: I1205 10:19:43.221286 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-catalog-content\") pod \"community-operators-2fbnj\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:43 crc kubenswrapper[4815]: I1205 10:19:43.221326 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jf2vw\" (UniqueName: \"kubernetes.io/projected/84f1fc5a-d663-4142-9ebf-ef103227ad90-kube-api-access-jf2vw\") pod \"community-operators-2fbnj\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:43 crc kubenswrapper[4815]: I1205 10:19:43.322635 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-utilities\") pod \"community-operators-2fbnj\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:43 crc kubenswrapper[4815]: I1205 10:19:43.322773 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-catalog-content\") pod \"community-operators-2fbnj\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:43 crc kubenswrapper[4815]: I1205 10:19:43.322827 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jf2vw\" (UniqueName: \"kubernetes.io/projected/84f1fc5a-d663-4142-9ebf-ef103227ad90-kube-api-access-jf2vw\") pod \"community-operators-2fbnj\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:43 crc kubenswrapper[4815]: I1205 10:19:43.323256 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-utilities\") pod \"community-operators-2fbnj\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:43 crc kubenswrapper[4815]: I1205 10:19:43.323391 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-catalog-content\") pod \"community-operators-2fbnj\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:43 crc kubenswrapper[4815]: I1205 10:19:43.345369 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jf2vw\" (UniqueName: \"kubernetes.io/projected/84f1fc5a-d663-4142-9ebf-ef103227ad90-kube-api-access-jf2vw\") pod \"community-operators-2fbnj\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:43 crc kubenswrapper[4815]: I1205 10:19:43.617647 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:44 crc kubenswrapper[4815]: I1205 10:19:44.514390 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-2fbnj"] Dec 05 10:19:45 crc kubenswrapper[4815]: I1205 10:19:45.019363 4815 generic.go:334] "Generic (PLEG): container finished" podID="84f1fc5a-d663-4142-9ebf-ef103227ad90" containerID="0044880c157baa49f6a8844bbdfd12fc495f5bd8cc0e69fd3d55ec4c5d505689" exitCode=0 Dec 05 10:19:45 crc kubenswrapper[4815]: I1205 10:19:45.019428 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fbnj" event={"ID":"84f1fc5a-d663-4142-9ebf-ef103227ad90","Type":"ContainerDied","Data":"0044880c157baa49f6a8844bbdfd12fc495f5bd8cc0e69fd3d55ec4c5d505689"} Dec 05 10:19:45 crc kubenswrapper[4815]: I1205 10:19:45.019696 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fbnj" event={"ID":"84f1fc5a-d663-4142-9ebf-ef103227ad90","Type":"ContainerStarted","Data":"ae3dd40c5e47004dc6e1f35a9bae501878d046a6d402635e9aa34ccdb43d35bf"} Dec 05 10:19:46 crc kubenswrapper[4815]: I1205 10:19:46.035144 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fbnj" event={"ID":"84f1fc5a-d663-4142-9ebf-ef103227ad90","Type":"ContainerStarted","Data":"da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5"} Dec 05 10:19:47 crc kubenswrapper[4815]: I1205 10:19:47.058314 4815 generic.go:334] "Generic (PLEG): container finished" podID="84f1fc5a-d663-4142-9ebf-ef103227ad90" containerID="da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5" exitCode=0 Dec 05 10:19:47 crc kubenswrapper[4815]: I1205 10:19:47.059008 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fbnj" event={"ID":"84f1fc5a-d663-4142-9ebf-ef103227ad90","Type":"ContainerDied","Data":"da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5"} Dec 05 10:19:48 crc kubenswrapper[4815]: I1205 10:19:48.075169 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fbnj" event={"ID":"84f1fc5a-d663-4142-9ebf-ef103227ad90","Type":"ContainerStarted","Data":"11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078"} Dec 05 10:19:48 crc kubenswrapper[4815]: I1205 10:19:48.107028 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-2fbnj" podStartSLOduration=3.666061489 podStartE2EDuration="6.106985459s" podCreationTimestamp="2025-12-05 10:19:42 +0000 UTC" firstStartedPulling="2025-12-05 10:19:45.021024914 +0000 UTC m=+4383.899631751" lastFinishedPulling="2025-12-05 10:19:47.461948884 +0000 UTC m=+4386.340555721" observedRunningTime="2025-12-05 10:19:48.097921752 +0000 UTC m=+4386.976528589" watchObservedRunningTime="2025-12-05 10:19:48.106985459 +0000 UTC m=+4386.985592296" Dec 05 10:19:50 crc kubenswrapper[4815]: I1205 10:19:50.192866 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:19:50 crc kubenswrapper[4815]: I1205 10:19:50.193423 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:19:50 crc kubenswrapper[4815]: I1205 10:19:50.193507 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 10:19:50 crc kubenswrapper[4815]: I1205 10:19:50.194327 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:19:50 crc kubenswrapper[4815]: I1205 10:19:50.194378 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" gracePeriod=600 Dec 05 10:19:50 crc kubenswrapper[4815]: E1205 10:19:50.832974 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:19:51 crc kubenswrapper[4815]: I1205 10:19:51.104111 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" exitCode=0 Dec 05 10:19:51 crc kubenswrapper[4815]: I1205 10:19:51.104161 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739"} Dec 05 10:19:51 crc kubenswrapper[4815]: I1205 10:19:51.104525 4815 scope.go:117] "RemoveContainer" containerID="b302353b8ed6e7609eaaca87ebca2fffa6b7c967de913ce1b3341b9b586d2603" Dec 05 10:19:51 crc kubenswrapper[4815]: I1205 10:19:51.105379 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:19:51 crc kubenswrapper[4815]: E1205 10:19:51.105849 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:19:53 crc kubenswrapper[4815]: I1205 10:19:53.702363 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:53 crc kubenswrapper[4815]: I1205 10:19:53.702428 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:53 crc kubenswrapper[4815]: I1205 10:19:53.759121 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:54 crc kubenswrapper[4815]: I1205 10:19:54.188742 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:54 crc kubenswrapper[4815]: I1205 10:19:54.254917 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2fbnj"] Dec 05 10:19:56 crc kubenswrapper[4815]: I1205 10:19:56.149462 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-2fbnj" podUID="84f1fc5a-d663-4142-9ebf-ef103227ad90" containerName="registry-server" containerID="cri-o://11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078" gracePeriod=2 Dec 05 10:19:56 crc kubenswrapper[4815]: I1205 10:19:56.637125 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:56 crc kubenswrapper[4815]: I1205 10:19:56.659544 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jf2vw\" (UniqueName: \"kubernetes.io/projected/84f1fc5a-d663-4142-9ebf-ef103227ad90-kube-api-access-jf2vw\") pod \"84f1fc5a-d663-4142-9ebf-ef103227ad90\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " Dec 05 10:19:56 crc kubenswrapper[4815]: I1205 10:19:56.659623 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-utilities\") pod \"84f1fc5a-d663-4142-9ebf-ef103227ad90\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " Dec 05 10:19:56 crc kubenswrapper[4815]: I1205 10:19:56.659703 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-catalog-content\") pod \"84f1fc5a-d663-4142-9ebf-ef103227ad90\" (UID: \"84f1fc5a-d663-4142-9ebf-ef103227ad90\") " Dec 05 10:19:56 crc kubenswrapper[4815]: I1205 10:19:56.665387 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-utilities" (OuterVolumeSpecName: "utilities") pod "84f1fc5a-d663-4142-9ebf-ef103227ad90" (UID: "84f1fc5a-d663-4142-9ebf-ef103227ad90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:19:56 crc kubenswrapper[4815]: I1205 10:19:56.679833 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84f1fc5a-d663-4142-9ebf-ef103227ad90-kube-api-access-jf2vw" (OuterVolumeSpecName: "kube-api-access-jf2vw") pod "84f1fc5a-d663-4142-9ebf-ef103227ad90" (UID: "84f1fc5a-d663-4142-9ebf-ef103227ad90"). InnerVolumeSpecName "kube-api-access-jf2vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:19:56 crc kubenswrapper[4815]: I1205 10:19:56.723429 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "84f1fc5a-d663-4142-9ebf-ef103227ad90" (UID: "84f1fc5a-d663-4142-9ebf-ef103227ad90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:19:56 crc kubenswrapper[4815]: I1205 10:19:56.762071 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jf2vw\" (UniqueName: \"kubernetes.io/projected/84f1fc5a-d663-4142-9ebf-ef103227ad90-kube-api-access-jf2vw\") on node \"crc\" DevicePath \"\"" Dec 05 10:19:56 crc kubenswrapper[4815]: I1205 10:19:56.762119 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:19:56 crc kubenswrapper[4815]: I1205 10:19:56.762131 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/84f1fc5a-d663-4142-9ebf-ef103227ad90-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.176475 4815 generic.go:334] "Generic (PLEG): container finished" podID="84f1fc5a-d663-4142-9ebf-ef103227ad90" containerID="11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078" exitCode=0 Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.176597 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fbnj" event={"ID":"84f1fc5a-d663-4142-9ebf-ef103227ad90","Type":"ContainerDied","Data":"11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078"} Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.176645 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-2fbnj" event={"ID":"84f1fc5a-d663-4142-9ebf-ef103227ad90","Type":"ContainerDied","Data":"ae3dd40c5e47004dc6e1f35a9bae501878d046a6d402635e9aa34ccdb43d35bf"} Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.176684 4815 scope.go:117] "RemoveContainer" containerID="11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078" Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.176707 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-2fbnj" Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.224073 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-2fbnj"] Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.227687 4815 scope.go:117] "RemoveContainer" containerID="da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5" Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.232642 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-2fbnj"] Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.251861 4815 scope.go:117] "RemoveContainer" containerID="0044880c157baa49f6a8844bbdfd12fc495f5bd8cc0e69fd3d55ec4c5d505689" Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.290951 4815 scope.go:117] "RemoveContainer" containerID="11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078" Dec 05 10:19:57 crc kubenswrapper[4815]: E1205 10:19:57.291397 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078\": container with ID starting with 11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078 not found: ID does not exist" containerID="11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078" Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.291438 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078"} err="failed to get container status \"11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078\": rpc error: code = NotFound desc = could not find container \"11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078\": container with ID starting with 11d8394e7d828fc98c0bfbbb90415830e1d7eb39c832c4f744fd2d18322d2078 not found: ID does not exist" Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.291465 4815 scope.go:117] "RemoveContainer" containerID="da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5" Dec 05 10:19:57 crc kubenswrapper[4815]: E1205 10:19:57.291764 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5\": container with ID starting with da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5 not found: ID does not exist" containerID="da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5" Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.291799 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5"} err="failed to get container status \"da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5\": rpc error: code = NotFound desc = could not find container \"da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5\": container with ID starting with da155c1100e54c0f9b02f79d63b2a5a48c538c9d70e91e9fdb281e5fadf8c4c5 not found: ID does not exist" Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.291823 4815 scope.go:117] "RemoveContainer" containerID="0044880c157baa49f6a8844bbdfd12fc495f5bd8cc0e69fd3d55ec4c5d505689" Dec 05 10:19:57 crc kubenswrapper[4815]: E1205 10:19:57.292075 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0044880c157baa49f6a8844bbdfd12fc495f5bd8cc0e69fd3d55ec4c5d505689\": container with ID starting with 0044880c157baa49f6a8844bbdfd12fc495f5bd8cc0e69fd3d55ec4c5d505689 not found: ID does not exist" containerID="0044880c157baa49f6a8844bbdfd12fc495f5bd8cc0e69fd3d55ec4c5d505689" Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.292130 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0044880c157baa49f6a8844bbdfd12fc495f5bd8cc0e69fd3d55ec4c5d505689"} err="failed to get container status \"0044880c157baa49f6a8844bbdfd12fc495f5bd8cc0e69fd3d55ec4c5d505689\": rpc error: code = NotFound desc = could not find container \"0044880c157baa49f6a8844bbdfd12fc495f5bd8cc0e69fd3d55ec4c5d505689\": container with ID starting with 0044880c157baa49f6a8844bbdfd12fc495f5bd8cc0e69fd3d55ec4c5d505689 not found: ID does not exist" Dec 05 10:19:57 crc kubenswrapper[4815]: I1205 10:19:57.430259 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84f1fc5a-d663-4142-9ebf-ef103227ad90" path="/var/lib/kubelet/pods/84f1fc5a-d663-4142-9ebf-ef103227ad90/volumes" Dec 05 10:20:05 crc kubenswrapper[4815]: I1205 10:20:05.419998 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:20:05 crc kubenswrapper[4815]: E1205 10:20:05.421651 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:20:18 crc kubenswrapper[4815]: I1205 10:20:18.418733 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:20:18 crc kubenswrapper[4815]: E1205 10:20:18.419420 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:20:32 crc kubenswrapper[4815]: I1205 10:20:32.419228 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:20:32 crc kubenswrapper[4815]: E1205 10:20:32.420016 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:20:44 crc kubenswrapper[4815]: I1205 10:20:44.418289 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:20:44 crc kubenswrapper[4815]: E1205 10:20:44.419067 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:20:57 crc kubenswrapper[4815]: I1205 10:20:57.418807 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:20:57 crc kubenswrapper[4815]: E1205 10:20:57.419612 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:21:09 crc kubenswrapper[4815]: I1205 10:21:09.419247 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:21:09 crc kubenswrapper[4815]: E1205 10:21:09.420175 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:21:21 crc kubenswrapper[4815]: I1205 10:21:21.427223 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:21:21 crc kubenswrapper[4815]: E1205 10:21:21.428227 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:21:35 crc kubenswrapper[4815]: I1205 10:21:35.418525 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:21:35 crc kubenswrapper[4815]: E1205 10:21:35.419277 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:21:47 crc kubenswrapper[4815]: I1205 10:21:47.418942 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:21:47 crc kubenswrapper[4815]: E1205 10:21:47.419874 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:21:59 crc kubenswrapper[4815]: I1205 10:21:59.418916 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:21:59 crc kubenswrapper[4815]: E1205 10:21:59.420199 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:22:13 crc kubenswrapper[4815]: I1205 10:22:13.418584 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:22:13 crc kubenswrapper[4815]: E1205 10:22:13.419548 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:22:14 crc kubenswrapper[4815]: I1205 10:22:14.853684 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-b6cnn" podUID="46274b62-2000-4c62-8688-af6b5fcb87dc" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.49:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 10:22:25 crc kubenswrapper[4815]: I1205 10:22:25.418682 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:22:25 crc kubenswrapper[4815]: E1205 10:22:25.419612 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:22:38 crc kubenswrapper[4815]: I1205 10:22:38.418893 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:22:38 crc kubenswrapper[4815]: E1205 10:22:38.419569 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:22:45 crc kubenswrapper[4815]: I1205 10:22:45.226833 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-57548d458d-9pflh" podUID="4abce90f-180f-4302-aecd-89bf5739757c" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 10:22:50 crc kubenswrapper[4815]: I1205 10:22:50.419316 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:22:50 crc kubenswrapper[4815]: E1205 10:22:50.421317 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:23:03 crc kubenswrapper[4815]: I1205 10:23:03.419647 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:23:03 crc kubenswrapper[4815]: E1205 10:23:03.420415 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:23:15 crc kubenswrapper[4815]: I1205 10:23:15.418818 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:23:15 crc kubenswrapper[4815]: E1205 10:23:15.419652 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:23:30 crc kubenswrapper[4815]: I1205 10:23:30.419325 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:23:30 crc kubenswrapper[4815]: E1205 10:23:30.420208 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:23:45 crc kubenswrapper[4815]: I1205 10:23:45.418976 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:23:45 crc kubenswrapper[4815]: E1205 10:23:45.419860 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:23:57 crc kubenswrapper[4815]: I1205 10:23:57.419974 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:23:57 crc kubenswrapper[4815]: E1205 10:23:57.421159 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:24:08 crc kubenswrapper[4815]: I1205 10:24:08.418580 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:24:08 crc kubenswrapper[4815]: E1205 10:24:08.419535 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:24:19 crc kubenswrapper[4815]: I1205 10:24:19.419767 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:24:19 crc kubenswrapper[4815]: E1205 10:24:19.421908 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:24:31 crc kubenswrapper[4815]: I1205 10:24:31.424113 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:24:31 crc kubenswrapper[4815]: E1205 10:24:31.424931 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:24:46 crc kubenswrapper[4815]: I1205 10:24:46.419265 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:24:46 crc kubenswrapper[4815]: E1205 10:24:46.420226 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:24:59 crc kubenswrapper[4815]: I1205 10:24:59.420162 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:25:00 crc kubenswrapper[4815]: I1205 10:25:00.264914 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"d34cc10d62df8ab9e8c19668427c9711c21dc528084d7e09c84a9817ac5c73f0"} Dec 05 10:26:17 crc kubenswrapper[4815]: I1205 10:26:17.049713 4815 generic.go:334] "Generic (PLEG): container finished" podID="b0ec9924-8223-4a07-806b-0ee6451a7021" containerID="2294af189294c56c3a263e2762c25a03ddca150493001b774200fbdeef9467e3" exitCode=0 Dec 05 10:26:17 crc kubenswrapper[4815]: I1205 10:26:17.050226 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"b0ec9924-8223-4a07-806b-0ee6451a7021","Type":"ContainerDied","Data":"2294af189294c56c3a263e2762c25a03ddca150493001b774200fbdeef9467e3"} Dec 05 10:26:18 crc kubenswrapper[4815]: I1205 10:26:18.939120 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.067206 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"b0ec9924-8223-4a07-806b-0ee6451a7021","Type":"ContainerDied","Data":"c70aa725f20c3733c9256756deba886a63e7b8e6d47cea8c8d11ebd6211b2421"} Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.067241 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c70aa725f20c3733c9256756deba886a63e7b8e6d47cea8c8d11ebd6211b2421" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.067282 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.109360 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config-secret\") pod \"b0ec9924-8223-4a07-806b-0ee6451a7021\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.109873 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config\") pod \"b0ec9924-8223-4a07-806b-0ee6451a7021\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.110028 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-config-data\") pod \"b0ec9924-8223-4a07-806b-0ee6451a7021\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.110588 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-temporary\") pod \"b0ec9924-8223-4a07-806b-0ee6451a7021\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.110666 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-workdir\") pod \"b0ec9924-8223-4a07-806b-0ee6451a7021\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.110695 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ca-certs\") pod \"b0ec9924-8223-4a07-806b-0ee6451a7021\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.110734 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"b0ec9924-8223-4a07-806b-0ee6451a7021\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.110760 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45lb8\" (UniqueName: \"kubernetes.io/projected/b0ec9924-8223-4a07-806b-0ee6451a7021-kube-api-access-45lb8\") pod \"b0ec9924-8223-4a07-806b-0ee6451a7021\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.110821 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ssh-key\") pod \"b0ec9924-8223-4a07-806b-0ee6451a7021\" (UID: \"b0ec9924-8223-4a07-806b-0ee6451a7021\") " Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.112216 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-config-data" (OuterVolumeSpecName: "config-data") pod "b0ec9924-8223-4a07-806b-0ee6451a7021" (UID: "b0ec9924-8223-4a07-806b-0ee6451a7021"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.112331 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "b0ec9924-8223-4a07-806b-0ee6451a7021" (UID: "b0ec9924-8223-4a07-806b-0ee6451a7021"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.127411 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "b0ec9924-8223-4a07-806b-0ee6451a7021" (UID: "b0ec9924-8223-4a07-806b-0ee6451a7021"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.130812 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "test-operator-logs") pod "b0ec9924-8223-4a07-806b-0ee6451a7021" (UID: "b0ec9924-8223-4a07-806b-0ee6451a7021"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.136854 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ec9924-8223-4a07-806b-0ee6451a7021-kube-api-access-45lb8" (OuterVolumeSpecName: "kube-api-access-45lb8") pod "b0ec9924-8223-4a07-806b-0ee6451a7021" (UID: "b0ec9924-8223-4a07-806b-0ee6451a7021"). InnerVolumeSpecName "kube-api-access-45lb8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.146649 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "b0ec9924-8223-4a07-806b-0ee6451a7021" (UID: "b0ec9924-8223-4a07-806b-0ee6451a7021"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.146127 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b0ec9924-8223-4a07-806b-0ee6451a7021" (UID: "b0ec9924-8223-4a07-806b-0ee6451a7021"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.148042 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "b0ec9924-8223-4a07-806b-0ee6451a7021" (UID: "b0ec9924-8223-4a07-806b-0ee6451a7021"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.174420 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "b0ec9924-8223-4a07-806b-0ee6451a7021" (UID: "b0ec9924-8223-4a07-806b-0ee6451a7021"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.212393 4815 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.212445 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45lb8\" (UniqueName: \"kubernetes.io/projected/b0ec9924-8223-4a07-806b-0ee6451a7021-kube-api-access-45lb8\") on node \"crc\" DevicePath \"\"" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.212461 4815 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.212474 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.212645 4815 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.212674 4815 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b0ec9924-8223-4a07-806b-0ee6451a7021-config-data\") on node \"crc\" DevicePath \"\"" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.212689 4815 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.212702 4815 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b0ec9924-8223-4a07-806b-0ee6451a7021-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.212719 4815 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b0ec9924-8223-4a07-806b-0ee6451a7021-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.239626 4815 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 05 10:26:19 crc kubenswrapper[4815]: I1205 10:26:19.314732 4815 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.070418 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 05 10:26:23 crc kubenswrapper[4815]: E1205 10:26:23.071091 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f1fc5a-d663-4142-9ebf-ef103227ad90" containerName="registry-server" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.071115 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f1fc5a-d663-4142-9ebf-ef103227ad90" containerName="registry-server" Dec 05 10:26:23 crc kubenswrapper[4815]: E1205 10:26:23.071138 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ec9924-8223-4a07-806b-0ee6451a7021" containerName="tempest-tests-tempest-tests-runner" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.071144 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ec9924-8223-4a07-806b-0ee6451a7021" containerName="tempest-tests-tempest-tests-runner" Dec 05 10:26:23 crc kubenswrapper[4815]: E1205 10:26:23.071155 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f1fc5a-d663-4142-9ebf-ef103227ad90" containerName="extract-content" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.071162 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f1fc5a-d663-4142-9ebf-ef103227ad90" containerName="extract-content" Dec 05 10:26:23 crc kubenswrapper[4815]: E1205 10:26:23.071193 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84f1fc5a-d663-4142-9ebf-ef103227ad90" containerName="extract-utilities" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.071199 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="84f1fc5a-d663-4142-9ebf-ef103227ad90" containerName="extract-utilities" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.071396 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="84f1fc5a-d663-4142-9ebf-ef103227ad90" containerName="registry-server" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.071419 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ec9924-8223-4a07-806b-0ee6451a7021" containerName="tempest-tests-tempest-tests-runner" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.072079 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.076651 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-r7768" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.080356 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.196647 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6fb2333c-6791-4164-a041-fec86a88cc06\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.196829 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qrt9\" (UniqueName: \"kubernetes.io/projected/6fb2333c-6791-4164-a041-fec86a88cc06-kube-api-access-2qrt9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6fb2333c-6791-4164-a041-fec86a88cc06\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.299275 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6fb2333c-6791-4164-a041-fec86a88cc06\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.299585 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qrt9\" (UniqueName: \"kubernetes.io/projected/6fb2333c-6791-4164-a041-fec86a88cc06-kube-api-access-2qrt9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6fb2333c-6791-4164-a041-fec86a88cc06\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.300514 4815 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6fb2333c-6791-4164-a041-fec86a88cc06\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.333316 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6fb2333c-6791-4164-a041-fec86a88cc06\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.336819 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qrt9\" (UniqueName: \"kubernetes.io/projected/6fb2333c-6791-4164-a041-fec86a88cc06-kube-api-access-2qrt9\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"6fb2333c-6791-4164-a041-fec86a88cc06\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.394139 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.965255 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 05 10:26:23 crc kubenswrapper[4815]: I1205 10:26:23.988174 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:26:24 crc kubenswrapper[4815]: I1205 10:26:24.115663 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"6fb2333c-6791-4164-a041-fec86a88cc06","Type":"ContainerStarted","Data":"1ddba5efb9537a615e4e6c2c729841a9b9fd5d578f73e77132bed83647e3190d"} Dec 05 10:26:26 crc kubenswrapper[4815]: I1205 10:26:26.138280 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"6fb2333c-6791-4164-a041-fec86a88cc06","Type":"ContainerStarted","Data":"1f9f8b4a92ec0b22a6460c7141c3e198a70c29ef0fe07a7d52a0e78bb1d57ea2"} Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.391179 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=27.326747829 podStartE2EDuration="28.391148739s" podCreationTimestamp="2025-12-05 10:26:23 +0000 UTC" firstStartedPulling="2025-12-05 10:26:23.98791918 +0000 UTC m=+4782.866526007" lastFinishedPulling="2025-12-05 10:26:25.05232006 +0000 UTC m=+4783.930926917" observedRunningTime="2025-12-05 10:26:26.166834252 +0000 UTC m=+4785.045441089" watchObservedRunningTime="2025-12-05 10:26:51.391148739 +0000 UTC m=+4810.269755576" Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.395663 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w2mx6/must-gather-gsjwp"] Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.399203 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/must-gather-gsjwp" Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.410525 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-w2mx6/must-gather-gsjwp"] Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.413085 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-w2mx6"/"openshift-service-ca.crt" Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.413328 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-w2mx6"/"default-dockercfg-ldgkk" Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.413815 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-w2mx6"/"kube-root-ca.crt" Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.602999 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-must-gather-output\") pod \"must-gather-gsjwp\" (UID: \"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2\") " pod="openshift-must-gather-w2mx6/must-gather-gsjwp" Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.603071 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlpxc\" (UniqueName: \"kubernetes.io/projected/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-kube-api-access-qlpxc\") pod \"must-gather-gsjwp\" (UID: \"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2\") " pod="openshift-must-gather-w2mx6/must-gather-gsjwp" Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.705313 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlpxc\" (UniqueName: \"kubernetes.io/projected/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-kube-api-access-qlpxc\") pod \"must-gather-gsjwp\" (UID: \"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2\") " pod="openshift-must-gather-w2mx6/must-gather-gsjwp" Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.705517 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-must-gather-output\") pod \"must-gather-gsjwp\" (UID: \"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2\") " pod="openshift-must-gather-w2mx6/must-gather-gsjwp" Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.705921 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-must-gather-output\") pod \"must-gather-gsjwp\" (UID: \"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2\") " pod="openshift-must-gather-w2mx6/must-gather-gsjwp" Dec 05 10:26:51 crc kubenswrapper[4815]: I1205 10:26:51.731922 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlpxc\" (UniqueName: \"kubernetes.io/projected/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-kube-api-access-qlpxc\") pod \"must-gather-gsjwp\" (UID: \"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2\") " pod="openshift-must-gather-w2mx6/must-gather-gsjwp" Dec 05 10:26:52 crc kubenswrapper[4815]: I1205 10:26:52.024298 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/must-gather-gsjwp" Dec 05 10:26:52 crc kubenswrapper[4815]: I1205 10:26:52.565891 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-w2mx6/must-gather-gsjwp"] Dec 05 10:26:53 crc kubenswrapper[4815]: I1205 10:26:53.400481 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w2mx6/must-gather-gsjwp" event={"ID":"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2","Type":"ContainerStarted","Data":"f8c55ef4f0ebf93d3c5ab5de3142c30582cd2b77e5c1ec9aa564328668e6a921"} Dec 05 10:26:59 crc kubenswrapper[4815]: I1205 10:26:59.964101 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w2mx6/must-gather-gsjwp" event={"ID":"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2","Type":"ContainerStarted","Data":"e6ef2f836c755fb6720d5d69ca50e83d05646c9e0badb80bc694ea53f50bc7f6"} Dec 05 10:26:59 crc kubenswrapper[4815]: I1205 10:26:59.964885 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w2mx6/must-gather-gsjwp" event={"ID":"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2","Type":"ContainerStarted","Data":"595cff59c7f177648979507968c0b00f42dc7874b2607fe12f56d5f62dfea3da"} Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:04.869840 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-jbdfv" podUID="563ef3a0-d97f-4b5b-afc4-378bdbbee19b" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:05.837555 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" podUID="c4cb3351-3a1f-4552-a820-841dea326932" containerName="ovnkube-controller" probeResult="failure" output="command timed out" Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:09.830282 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="078b665f-5e54-4d83-966b-3684d06ea320" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:09.830279 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" podUID="c4cb3351-3a1f-4552-a820-841dea326932" containerName="sbdb" probeResult="failure" output="command timed out" Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:09.830292 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-7q4vr" podUID="c4cb3351-3a1f-4552-a820-841dea326932" containerName="nbdb" probeResult="failure" output="command timed out" Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:10.078846 4815 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.443332312s: [/var/lib/containers/storage/overlay/da716153f13d4b04d09b4d6c881260724317a7bc5114066278e939b544b9ab53/diff /var/log/pods/openstack_rabbitmq-cell1-server-0_db8bc583-6bcf-40c7-875b-d5c9544af543/rabbitmq/0.log]; will not log again for this container unless duration exceeds 2s Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:10.080792 4815 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.826899069s: [/var/lib/containers/storage/overlay/7cf642911b66acec125f021696f4283c434c44728a309e1f96f6119f8a6da38d/diff /var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-bm8kq_3665adde-25ad-4261-be50-ef4e21726b7b/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:10.089650 4815 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.781798693s: [/var/lib/containers/storage/overlay/8fb191875a5a1de4d58bc2f3cad336ab576037f86bf877ff01e00a363f219a8a/diff ]; will not log again for this container unless duration exceeds 2s Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:10.156412 4815 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.521663988s: [/var/lib/containers/storage/overlay/b1d5c6681317e13313bfb37c7453cecf0555c0ff3f937ee5c9c93e85785d415f/diff /var/log/pods/openstack_rabbitmq-server-0_d0d7070e-5d74-4d40-9fe9-b42c60585002/rabbitmq/0.log]; will not log again for this container unless duration exceeds 2s Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:10.157481 4815 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.139447641s: [/var/lib/containers/storage/overlay/6ad9159060994d880aed0005df01d95d154f2ec1f70a654f258e80ba74ba9276/diff /var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-26tt9_cafb574b-a659-41e6-89d5-c02851274f73/manager/0.log]; will not log again for this container unless duration exceeds 2s Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:10.157571 4815 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.601085083s: [/var/lib/containers/storage/overlay/690120b8cebf52da70f6c1f778111e2d45a26869683c159049adb12992818eee/diff /var/log/pods/openshift-authentication_oauth-openshift-574dcf5686-rfnw4_90c29ff2-2ef3-4c5b-ba40-623d030ba101/oauth-openshift/0.log]; will not log again for this container unless duration exceeds 2s Dec 05 10:27:10 crc kubenswrapper[4815]: I1205 10:27:10.826948 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1" containerName="galera" probeResult="failure" output="command timed out" Dec 05 10:27:12 crc kubenswrapper[4815]: I1205 10:27:12.664629 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-w2mx6/must-gather-gsjwp" podStartSLOduration=15.641351823 podStartE2EDuration="21.664604472s" podCreationTimestamp="2025-12-05 10:26:51 +0000 UTC" firstStartedPulling="2025-12-05 10:26:52.572698888 +0000 UTC m=+4811.451305725" lastFinishedPulling="2025-12-05 10:26:58.595951537 +0000 UTC m=+4817.474558374" observedRunningTime="2025-12-05 10:26:59.989505744 +0000 UTC m=+4818.868112611" watchObservedRunningTime="2025-12-05 10:27:12.664604472 +0000 UTC m=+4831.543211309" Dec 05 10:27:12 crc kubenswrapper[4815]: I1205 10:27:12.670702 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w2mx6/crc-debug-vt2h2"] Dec 05 10:27:12 crc kubenswrapper[4815]: I1205 10:27:12.672154 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" Dec 05 10:27:12 crc kubenswrapper[4815]: I1205 10:27:12.753645 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfz62\" (UniqueName: \"kubernetes.io/projected/7c32a0ff-118b-482f-9e73-15ece3db9f88-kube-api-access-dfz62\") pod \"crc-debug-vt2h2\" (UID: \"7c32a0ff-118b-482f-9e73-15ece3db9f88\") " pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" Dec 05 10:27:12 crc kubenswrapper[4815]: I1205 10:27:12.754057 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c32a0ff-118b-482f-9e73-15ece3db9f88-host\") pod \"crc-debug-vt2h2\" (UID: \"7c32a0ff-118b-482f-9e73-15ece3db9f88\") " pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" Dec 05 10:27:12 crc kubenswrapper[4815]: I1205 10:27:12.855814 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfz62\" (UniqueName: \"kubernetes.io/projected/7c32a0ff-118b-482f-9e73-15ece3db9f88-kube-api-access-dfz62\") pod \"crc-debug-vt2h2\" (UID: \"7c32a0ff-118b-482f-9e73-15ece3db9f88\") " pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" Dec 05 10:27:12 crc kubenswrapper[4815]: I1205 10:27:12.855902 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c32a0ff-118b-482f-9e73-15ece3db9f88-host\") pod \"crc-debug-vt2h2\" (UID: \"7c32a0ff-118b-482f-9e73-15ece3db9f88\") " pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" Dec 05 10:27:12 crc kubenswrapper[4815]: I1205 10:27:12.862582 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c32a0ff-118b-482f-9e73-15ece3db9f88-host\") pod \"crc-debug-vt2h2\" (UID: \"7c32a0ff-118b-482f-9e73-15ece3db9f88\") " pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" Dec 05 10:27:12 crc kubenswrapper[4815]: I1205 10:27:12.895128 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfz62\" (UniqueName: \"kubernetes.io/projected/7c32a0ff-118b-482f-9e73-15ece3db9f88-kube-api-access-dfz62\") pod \"crc-debug-vt2h2\" (UID: \"7c32a0ff-118b-482f-9e73-15ece3db9f88\") " pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" Dec 05 10:27:12 crc kubenswrapper[4815]: I1205 10:27:12.993268 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" Dec 05 10:27:14 crc kubenswrapper[4815]: I1205 10:27:14.147926 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" event={"ID":"7c32a0ff-118b-482f-9e73-15ece3db9f88","Type":"ContainerStarted","Data":"ab0bfc35b268c388beea787aeb4285666c45568c95a86cafade190bc16654519"} Dec 05 10:27:20 crc kubenswrapper[4815]: I1205 10:27:20.192673 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:27:20 crc kubenswrapper[4815]: I1205 10:27:20.193235 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:27:24 crc kubenswrapper[4815]: I1205 10:27:24.872668 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-jbdfv" podUID="563ef3a0-d97f-4b5b-afc4-378bdbbee19b" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 10:27:30 crc kubenswrapper[4815]: E1205 10:27:30.782914 4815 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Dec 05 10:27:30 crc kubenswrapper[4815]: E1205 10:27:30.784538 4815 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dfz62,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-vt2h2_openshift-must-gather-w2mx6(7c32a0ff-118b-482f-9e73-15ece3db9f88): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 05 10:27:30 crc kubenswrapper[4815]: E1205 10:27:30.785736 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" podUID="7c32a0ff-118b-482f-9e73-15ece3db9f88" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.330405 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4ccx9"] Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.333766 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:31 crc kubenswrapper[4815]: E1205 10:27:31.401233 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" podUID="7c32a0ff-118b-482f-9e73-15ece3db9f88" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.412817 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4ccx9"] Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.429813 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-utilities\") pod \"certified-operators-4ccx9\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.429975 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcc7c\" (UniqueName: \"kubernetes.io/projected/1c743ae0-1add-469e-8d96-f6183f1b6404-kube-api-access-qcc7c\") pod \"certified-operators-4ccx9\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.430037 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-catalog-content\") pod \"certified-operators-4ccx9\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.523132 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zl22b"] Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.529560 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.547668 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcc7c\" (UniqueName: \"kubernetes.io/projected/1c743ae0-1add-469e-8d96-f6183f1b6404-kube-api-access-qcc7c\") pod \"certified-operators-4ccx9\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.548387 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-catalog-content\") pod \"certified-operators-4ccx9\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.549093 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-catalog-content\") pod \"certified-operators-4ccx9\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.549579 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-utilities\") pod \"certified-operators-4ccx9\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.551920 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-utilities\") pod \"certified-operators-4ccx9\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.552336 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zl22b"] Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.589025 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcc7c\" (UniqueName: \"kubernetes.io/projected/1c743ae0-1add-469e-8d96-f6183f1b6404-kube-api-access-qcc7c\") pod \"certified-operators-4ccx9\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.652097 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-catalog-content\") pod \"redhat-marketplace-zl22b\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.652443 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-utilities\") pod \"redhat-marketplace-zl22b\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.652649 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t7jx9\" (UniqueName: \"kubernetes.io/projected/39a4c4a1-9df9-42e1-9f76-604380617157-kube-api-access-t7jx9\") pod \"redhat-marketplace-zl22b\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.660009 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.754648 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-utilities\") pod \"redhat-marketplace-zl22b\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.754798 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t7jx9\" (UniqueName: \"kubernetes.io/projected/39a4c4a1-9df9-42e1-9f76-604380617157-kube-api-access-t7jx9\") pod \"redhat-marketplace-zl22b\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.754887 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-catalog-content\") pod \"redhat-marketplace-zl22b\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.755348 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-catalog-content\") pod \"redhat-marketplace-zl22b\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.755348 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-utilities\") pod \"redhat-marketplace-zl22b\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.778178 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t7jx9\" (UniqueName: \"kubernetes.io/projected/39a4c4a1-9df9-42e1-9f76-604380617157-kube-api-access-t7jx9\") pod \"redhat-marketplace-zl22b\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:31 crc kubenswrapper[4815]: I1205 10:27:31.850387 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:32 crc kubenswrapper[4815]: I1205 10:27:32.423588 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4ccx9"] Dec 05 10:27:32 crc kubenswrapper[4815]: I1205 10:27:32.543937 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zl22b"] Dec 05 10:27:33 crc kubenswrapper[4815]: I1205 10:27:33.436619 4815 generic.go:334] "Generic (PLEG): container finished" podID="39a4c4a1-9df9-42e1-9f76-604380617157" containerID="0c9972be46f5bf291c03fff85b1ef4fe6812616c7239504a0d8e5af1532978ad" exitCode=0 Dec 05 10:27:33 crc kubenswrapper[4815]: I1205 10:27:33.436629 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zl22b" event={"ID":"39a4c4a1-9df9-42e1-9f76-604380617157","Type":"ContainerDied","Data":"0c9972be46f5bf291c03fff85b1ef4fe6812616c7239504a0d8e5af1532978ad"} Dec 05 10:27:33 crc kubenswrapper[4815]: I1205 10:27:33.437339 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zl22b" event={"ID":"39a4c4a1-9df9-42e1-9f76-604380617157","Type":"ContainerStarted","Data":"2cdd3317f27ff8c93cc4190e1cac5b545b882f9b3e8f36390fa41ca5ed8cfa62"} Dec 05 10:27:33 crc kubenswrapper[4815]: I1205 10:27:33.439652 4815 generic.go:334] "Generic (PLEG): container finished" podID="1c743ae0-1add-469e-8d96-f6183f1b6404" containerID="6f6ad4fa3ed36c0d72c02204da595c93e9b78921a0bd5648b1361054f10ffb59" exitCode=0 Dec 05 10:27:33 crc kubenswrapper[4815]: I1205 10:27:33.439719 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ccx9" event={"ID":"1c743ae0-1add-469e-8d96-f6183f1b6404","Type":"ContainerDied","Data":"6f6ad4fa3ed36c0d72c02204da595c93e9b78921a0bd5648b1361054f10ffb59"} Dec 05 10:27:33 crc kubenswrapper[4815]: I1205 10:27:33.439862 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ccx9" event={"ID":"1c743ae0-1add-469e-8d96-f6183f1b6404","Type":"ContainerStarted","Data":"672e109123657fd400530829f0e811ba2feafeec6848d04855913d2ec01a5abc"} Dec 05 10:27:34 crc kubenswrapper[4815]: I1205 10:27:34.450304 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zl22b" event={"ID":"39a4c4a1-9df9-42e1-9f76-604380617157","Type":"ContainerStarted","Data":"521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897"} Dec 05 10:27:35 crc kubenswrapper[4815]: I1205 10:27:35.461737 4815 generic.go:334] "Generic (PLEG): container finished" podID="39a4c4a1-9df9-42e1-9f76-604380617157" containerID="521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897" exitCode=0 Dec 05 10:27:35 crc kubenswrapper[4815]: I1205 10:27:35.461919 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zl22b" event={"ID":"39a4c4a1-9df9-42e1-9f76-604380617157","Type":"ContainerDied","Data":"521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897"} Dec 05 10:27:35 crc kubenswrapper[4815]: I1205 10:27:35.464170 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ccx9" event={"ID":"1c743ae0-1add-469e-8d96-f6183f1b6404","Type":"ContainerStarted","Data":"b9c39accfa11678af00a232ec0a62ebd2934f230b428b87b2aff0bc4c9edfab8"} Dec 05 10:27:37 crc kubenswrapper[4815]: I1205 10:27:37.580169 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zl22b" event={"ID":"39a4c4a1-9df9-42e1-9f76-604380617157","Type":"ContainerStarted","Data":"81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe"} Dec 05 10:27:37 crc kubenswrapper[4815]: I1205 10:27:37.582870 4815 generic.go:334] "Generic (PLEG): container finished" podID="1c743ae0-1add-469e-8d96-f6183f1b6404" containerID="b9c39accfa11678af00a232ec0a62ebd2934f230b428b87b2aff0bc4c9edfab8" exitCode=0 Dec 05 10:27:37 crc kubenswrapper[4815]: I1205 10:27:37.582921 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ccx9" event={"ID":"1c743ae0-1add-469e-8d96-f6183f1b6404","Type":"ContainerDied","Data":"b9c39accfa11678af00a232ec0a62ebd2934f230b428b87b2aff0bc4c9edfab8"} Dec 05 10:27:37 crc kubenswrapper[4815]: I1205 10:27:37.607108 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zl22b" podStartSLOduration=3.789283856 podStartE2EDuration="6.607087497s" podCreationTimestamp="2025-12-05 10:27:31 +0000 UTC" firstStartedPulling="2025-12-05 10:27:33.440392975 +0000 UTC m=+4852.318999812" lastFinishedPulling="2025-12-05 10:27:36.258196616 +0000 UTC m=+4855.136803453" observedRunningTime="2025-12-05 10:27:37.603481959 +0000 UTC m=+4856.482088816" watchObservedRunningTime="2025-12-05 10:27:37.607087497 +0000 UTC m=+4856.485694354" Dec 05 10:27:38 crc kubenswrapper[4815]: I1205 10:27:38.598327 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ccx9" event={"ID":"1c743ae0-1add-469e-8d96-f6183f1b6404","Type":"ContainerStarted","Data":"e87bbcc2cd6b6156e25ad1cf8eaf9cbe0090651fcfb25c600710d1b6760ccbd3"} Dec 05 10:27:38 crc kubenswrapper[4815]: I1205 10:27:38.630852 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4ccx9" podStartSLOduration=2.998805585 podStartE2EDuration="7.630826478s" podCreationTimestamp="2025-12-05 10:27:31 +0000 UTC" firstStartedPulling="2025-12-05 10:27:33.443414418 +0000 UTC m=+4852.322021265" lastFinishedPulling="2025-12-05 10:27:38.075435331 +0000 UTC m=+4856.954042158" observedRunningTime="2025-12-05 10:27:38.628224936 +0000 UTC m=+4857.506831773" watchObservedRunningTime="2025-12-05 10:27:38.630826478 +0000 UTC m=+4857.509433315" Dec 05 10:27:41 crc kubenswrapper[4815]: I1205 10:27:41.660359 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:41 crc kubenswrapper[4815]: I1205 10:27:41.661907 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:41 crc kubenswrapper[4815]: I1205 10:27:41.709858 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:41 crc kubenswrapper[4815]: I1205 10:27:41.851523 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:41 crc kubenswrapper[4815]: I1205 10:27:41.851585 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:41 crc kubenswrapper[4815]: I1205 10:27:41.923269 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:42 crc kubenswrapper[4815]: I1205 10:27:42.705210 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:43 crc kubenswrapper[4815]: I1205 10:27:43.510523 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zl22b"] Dec 05 10:27:43 crc kubenswrapper[4815]: I1205 10:27:43.696239 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:44 crc kubenswrapper[4815]: I1205 10:27:44.655427 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zl22b" podUID="39a4c4a1-9df9-42e1-9f76-604380617157" containerName="registry-server" containerID="cri-o://81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe" gracePeriod=2 Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.206817 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.283544 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-utilities\") pod \"39a4c4a1-9df9-42e1-9f76-604380617157\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.283903 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-catalog-content\") pod \"39a4c4a1-9df9-42e1-9f76-604380617157\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.284065 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t7jx9\" (UniqueName: \"kubernetes.io/projected/39a4c4a1-9df9-42e1-9f76-604380617157-kube-api-access-t7jx9\") pod \"39a4c4a1-9df9-42e1-9f76-604380617157\" (UID: \"39a4c4a1-9df9-42e1-9f76-604380617157\") " Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.284941 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-utilities" (OuterVolumeSpecName: "utilities") pod "39a4c4a1-9df9-42e1-9f76-604380617157" (UID: "39a4c4a1-9df9-42e1-9f76-604380617157"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.290823 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39a4c4a1-9df9-42e1-9f76-604380617157-kube-api-access-t7jx9" (OuterVolumeSpecName: "kube-api-access-t7jx9") pod "39a4c4a1-9df9-42e1-9f76-604380617157" (UID: "39a4c4a1-9df9-42e1-9f76-604380617157"). InnerVolumeSpecName "kube-api-access-t7jx9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.313887 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "39a4c4a1-9df9-42e1-9f76-604380617157" (UID: "39a4c4a1-9df9-42e1-9f76-604380617157"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.403988 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.404024 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/39a4c4a1-9df9-42e1-9f76-604380617157-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.404035 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t7jx9\" (UniqueName: \"kubernetes.io/projected/39a4c4a1-9df9-42e1-9f76-604380617157-kube-api-access-t7jx9\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.664020 4815 generic.go:334] "Generic (PLEG): container finished" podID="39a4c4a1-9df9-42e1-9f76-604380617157" containerID="81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe" exitCode=0 Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.664061 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zl22b" event={"ID":"39a4c4a1-9df9-42e1-9f76-604380617157","Type":"ContainerDied","Data":"81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe"} Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.664086 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zl22b" event={"ID":"39a4c4a1-9df9-42e1-9f76-604380617157","Type":"ContainerDied","Data":"2cdd3317f27ff8c93cc4190e1cac5b545b882f9b3e8f36390fa41ca5ed8cfa62"} Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.664113 4815 scope.go:117] "RemoveContainer" containerID="81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.664233 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zl22b" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.693892 4815 scope.go:117] "RemoveContainer" containerID="521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.716330 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zl22b"] Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.720289 4815 scope.go:117] "RemoveContainer" containerID="0c9972be46f5bf291c03fff85b1ef4fe6812616c7239504a0d8e5af1532978ad" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.730482 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zl22b"] Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.795022 4815 scope.go:117] "RemoveContainer" containerID="81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe" Dec 05 10:27:45 crc kubenswrapper[4815]: E1205 10:27:45.795547 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe\": container with ID starting with 81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe not found: ID does not exist" containerID="81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.795586 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe"} err="failed to get container status \"81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe\": rpc error: code = NotFound desc = could not find container \"81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe\": container with ID starting with 81ea8a568e1156c02d4876d5bd785bed51193fad32b0a56a7606c7e836aa87fe not found: ID does not exist" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.795845 4815 scope.go:117] "RemoveContainer" containerID="521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897" Dec 05 10:27:45 crc kubenswrapper[4815]: E1205 10:27:45.796093 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897\": container with ID starting with 521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897 not found: ID does not exist" containerID="521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.796122 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897"} err="failed to get container status \"521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897\": rpc error: code = NotFound desc = could not find container \"521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897\": container with ID starting with 521d2c347ff85c44d3beecb00e4eb2217e6944f6eaff0561b101295e634ac897 not found: ID does not exist" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.796146 4815 scope.go:117] "RemoveContainer" containerID="0c9972be46f5bf291c03fff85b1ef4fe6812616c7239504a0d8e5af1532978ad" Dec 05 10:27:45 crc kubenswrapper[4815]: E1205 10:27:45.796378 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c9972be46f5bf291c03fff85b1ef4fe6812616c7239504a0d8e5af1532978ad\": container with ID starting with 0c9972be46f5bf291c03fff85b1ef4fe6812616c7239504a0d8e5af1532978ad not found: ID does not exist" containerID="0c9972be46f5bf291c03fff85b1ef4fe6812616c7239504a0d8e5af1532978ad" Dec 05 10:27:45 crc kubenswrapper[4815]: I1205 10:27:45.796422 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c9972be46f5bf291c03fff85b1ef4fe6812616c7239504a0d8e5af1532978ad"} err="failed to get container status \"0c9972be46f5bf291c03fff85b1ef4fe6812616c7239504a0d8e5af1532978ad\": rpc error: code = NotFound desc = could not find container \"0c9972be46f5bf291c03fff85b1ef4fe6812616c7239504a0d8e5af1532978ad\": container with ID starting with 0c9972be46f5bf291c03fff85b1ef4fe6812616c7239504a0d8e5af1532978ad not found: ID does not exist" Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.119411 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4ccx9"] Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.120412 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4ccx9" podUID="1c743ae0-1add-469e-8d96-f6183f1b6404" containerName="registry-server" containerID="cri-o://e87bbcc2cd6b6156e25ad1cf8eaf9cbe0090651fcfb25c600710d1b6760ccbd3" gracePeriod=2 Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.683482 4815 generic.go:334] "Generic (PLEG): container finished" podID="1c743ae0-1add-469e-8d96-f6183f1b6404" containerID="e87bbcc2cd6b6156e25ad1cf8eaf9cbe0090651fcfb25c600710d1b6760ccbd3" exitCode=0 Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.684140 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ccx9" event={"ID":"1c743ae0-1add-469e-8d96-f6183f1b6404","Type":"ContainerDied","Data":"e87bbcc2cd6b6156e25ad1cf8eaf9cbe0090651fcfb25c600710d1b6760ccbd3"} Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.684193 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4ccx9" event={"ID":"1c743ae0-1add-469e-8d96-f6183f1b6404","Type":"ContainerDied","Data":"672e109123657fd400530829f0e811ba2feafeec6848d04855913d2ec01a5abc"} Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.684206 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="672e109123657fd400530829f0e811ba2feafeec6848d04855913d2ec01a5abc" Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.686785 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" event={"ID":"7c32a0ff-118b-482f-9e73-15ece3db9f88","Type":"ContainerStarted","Data":"4ff58c55e0ac50c928ea60b59ec9724ba9f03cf007331dc37a3ca34b3973f952"} Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.698911 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.714198 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" podStartSLOduration=2.490617682 podStartE2EDuration="34.714175296s" podCreationTimestamp="2025-12-05 10:27:12 +0000 UTC" firstStartedPulling="2025-12-05 10:27:13.642264275 +0000 UTC m=+4832.520871102" lastFinishedPulling="2025-12-05 10:27:45.865821879 +0000 UTC m=+4864.744428716" observedRunningTime="2025-12-05 10:27:46.703195616 +0000 UTC m=+4865.581802463" watchObservedRunningTime="2025-12-05 10:27:46.714175296 +0000 UTC m=+4865.592782133" Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.741407 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcc7c\" (UniqueName: \"kubernetes.io/projected/1c743ae0-1add-469e-8d96-f6183f1b6404-kube-api-access-qcc7c\") pod \"1c743ae0-1add-469e-8d96-f6183f1b6404\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.741594 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-utilities\") pod \"1c743ae0-1add-469e-8d96-f6183f1b6404\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.742427 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-utilities" (OuterVolumeSpecName: "utilities") pod "1c743ae0-1add-469e-8d96-f6183f1b6404" (UID: "1c743ae0-1add-469e-8d96-f6183f1b6404"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.742508 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-catalog-content\") pod \"1c743ae0-1add-469e-8d96-f6183f1b6404\" (UID: \"1c743ae0-1add-469e-8d96-f6183f1b6404\") " Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.743387 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.760154 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c743ae0-1add-469e-8d96-f6183f1b6404-kube-api-access-qcc7c" (OuterVolumeSpecName: "kube-api-access-qcc7c") pod "1c743ae0-1add-469e-8d96-f6183f1b6404" (UID: "1c743ae0-1add-469e-8d96-f6183f1b6404"). InnerVolumeSpecName "kube-api-access-qcc7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.816731 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c743ae0-1add-469e-8d96-f6183f1b6404" (UID: "1c743ae0-1add-469e-8d96-f6183f1b6404"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.845477 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcc7c\" (UniqueName: \"kubernetes.io/projected/1c743ae0-1add-469e-8d96-f6183f1b6404-kube-api-access-qcc7c\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:46 crc kubenswrapper[4815]: I1205 10:27:46.845564 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c743ae0-1add-469e-8d96-f6183f1b6404-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:27:47 crc kubenswrapper[4815]: I1205 10:27:47.433050 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39a4c4a1-9df9-42e1-9f76-604380617157" path="/var/lib/kubelet/pods/39a4c4a1-9df9-42e1-9f76-604380617157/volumes" Dec 05 10:27:47 crc kubenswrapper[4815]: I1205 10:27:47.700259 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4ccx9" Dec 05 10:27:47 crc kubenswrapper[4815]: I1205 10:27:47.730435 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4ccx9"] Dec 05 10:27:47 crc kubenswrapper[4815]: I1205 10:27:47.741308 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4ccx9"] Dec 05 10:27:49 crc kubenswrapper[4815]: I1205 10:27:49.432656 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c743ae0-1add-469e-8d96-f6183f1b6404" path="/var/lib/kubelet/pods/1c743ae0-1add-469e-8d96-f6183f1b6404/volumes" Dec 05 10:27:50 crc kubenswrapper[4815]: I1205 10:27:50.192760 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:27:50 crc kubenswrapper[4815]: I1205 10:27:50.192816 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:28:01 crc kubenswrapper[4815]: I1205 10:28:01.264662 4815 patch_prober.go:28] interesting pod/console-operator-58897d9998-4mvlm container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.22:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 10:28:01 crc kubenswrapper[4815]: I1205 10:28:01.265177 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-4mvlm" podUID="a20f4d0e-9b06-4bc0-a532-a104392a18dd" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.22:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:28:20 crc kubenswrapper[4815]: I1205 10:28:20.192325 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:28:20 crc kubenswrapper[4815]: I1205 10:28:20.192928 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:28:20 crc kubenswrapper[4815]: I1205 10:28:20.192979 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 10:28:20 crc kubenswrapper[4815]: I1205 10:28:20.193654 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d34cc10d62df8ab9e8c19668427c9711c21dc528084d7e09c84a9817ac5c73f0"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:28:20 crc kubenswrapper[4815]: I1205 10:28:20.193711 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://d34cc10d62df8ab9e8c19668427c9711c21dc528084d7e09c84a9817ac5c73f0" gracePeriod=600 Dec 05 10:28:21 crc kubenswrapper[4815]: I1205 10:28:21.057548 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="d34cc10d62df8ab9e8c19668427c9711c21dc528084d7e09c84a9817ac5c73f0" exitCode=0 Dec 05 10:28:21 crc kubenswrapper[4815]: I1205 10:28:21.057612 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"d34cc10d62df8ab9e8c19668427c9711c21dc528084d7e09c84a9817ac5c73f0"} Dec 05 10:28:21 crc kubenswrapper[4815]: I1205 10:28:21.057694 4815 scope.go:117] "RemoveContainer" containerID="3714acde175698d03937a8439b18a3fdd07e01562764681b94cd58bdd1a00739" Dec 05 10:28:23 crc kubenswrapper[4815]: I1205 10:28:23.078975 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f"} Dec 05 10:28:26 crc kubenswrapper[4815]: I1205 10:28:26.198427 4815 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.013399807s: [/var/lib/containers/storage/overlay/d8587cf118e3836b75853d8f58fa6280112ff24d66b0aa171ea25f296c1a8848/diff /var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_6fb2333c-6791-4164-a041-fec86a88cc06/test-operator-logs-container/0.log]; will not log again for this container unless duration exceeds 2s Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.853739 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pcptj"] Dec 05 10:28:38 crc kubenswrapper[4815]: E1205 10:28:38.854732 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39a4c4a1-9df9-42e1-9f76-604380617157" containerName="extract-utilities" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.854772 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="39a4c4a1-9df9-42e1-9f76-604380617157" containerName="extract-utilities" Dec 05 10:28:38 crc kubenswrapper[4815]: E1205 10:28:38.854816 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39a4c4a1-9df9-42e1-9f76-604380617157" containerName="extract-content" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.854825 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="39a4c4a1-9df9-42e1-9f76-604380617157" containerName="extract-content" Dec 05 10:28:38 crc kubenswrapper[4815]: E1205 10:28:38.854838 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c743ae0-1add-469e-8d96-f6183f1b6404" containerName="extract-content" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.854846 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c743ae0-1add-469e-8d96-f6183f1b6404" containerName="extract-content" Dec 05 10:28:38 crc kubenswrapper[4815]: E1205 10:28:38.854867 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c743ae0-1add-469e-8d96-f6183f1b6404" containerName="registry-server" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.854875 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c743ae0-1add-469e-8d96-f6183f1b6404" containerName="registry-server" Dec 05 10:28:38 crc kubenswrapper[4815]: E1205 10:28:38.854891 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39a4c4a1-9df9-42e1-9f76-604380617157" containerName="registry-server" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.854899 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="39a4c4a1-9df9-42e1-9f76-604380617157" containerName="registry-server" Dec 05 10:28:38 crc kubenswrapper[4815]: E1205 10:28:38.854918 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c743ae0-1add-469e-8d96-f6183f1b6404" containerName="extract-utilities" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.854926 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c743ae0-1add-469e-8d96-f6183f1b6404" containerName="extract-utilities" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.855190 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="39a4c4a1-9df9-42e1-9f76-604380617157" containerName="registry-server" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.855217 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c743ae0-1add-469e-8d96-f6183f1b6404" containerName="registry-server" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.857425 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.893638 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pcptj"] Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.923032 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txsr6\" (UniqueName: \"kubernetes.io/projected/1da3a037-cf60-474f-be25-62e9601ee7bc-kube-api-access-txsr6\") pod \"redhat-operators-pcptj\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.923133 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-catalog-content\") pod \"redhat-operators-pcptj\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:38 crc kubenswrapper[4815]: I1205 10:28:38.923225 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-utilities\") pod \"redhat-operators-pcptj\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:39 crc kubenswrapper[4815]: I1205 10:28:39.024901 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txsr6\" (UniqueName: \"kubernetes.io/projected/1da3a037-cf60-474f-be25-62e9601ee7bc-kube-api-access-txsr6\") pod \"redhat-operators-pcptj\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:39 crc kubenswrapper[4815]: I1205 10:28:39.025017 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-catalog-content\") pod \"redhat-operators-pcptj\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:39 crc kubenswrapper[4815]: I1205 10:28:39.025128 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-utilities\") pod \"redhat-operators-pcptj\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:39 crc kubenswrapper[4815]: I1205 10:28:39.025970 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-utilities\") pod \"redhat-operators-pcptj\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:39 crc kubenswrapper[4815]: I1205 10:28:39.025980 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-catalog-content\") pod \"redhat-operators-pcptj\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:39 crc kubenswrapper[4815]: I1205 10:28:39.054552 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txsr6\" (UniqueName: \"kubernetes.io/projected/1da3a037-cf60-474f-be25-62e9601ee7bc-kube-api-access-txsr6\") pod \"redhat-operators-pcptj\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:39 crc kubenswrapper[4815]: I1205 10:28:39.193921 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:39 crc kubenswrapper[4815]: I1205 10:28:39.879152 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pcptj"] Dec 05 10:28:40 crc kubenswrapper[4815]: I1205 10:28:40.229603 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcptj" event={"ID":"1da3a037-cf60-474f-be25-62e9601ee7bc","Type":"ContainerStarted","Data":"b014e9932bae82d1ee5a5dd7d04fcb28f7a0bd4fbe405bdbd309889e0e42ffef"} Dec 05 10:28:42 crc kubenswrapper[4815]: I1205 10:28:42.253598 4815 generic.go:334] "Generic (PLEG): container finished" podID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerID="fed48841608062541922634baff7c538ecef5ef6786b478fd45588f0a2296dfe" exitCode=0 Dec 05 10:28:42 crc kubenswrapper[4815]: I1205 10:28:42.254160 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcptj" event={"ID":"1da3a037-cf60-474f-be25-62e9601ee7bc","Type":"ContainerDied","Data":"fed48841608062541922634baff7c538ecef5ef6786b478fd45588f0a2296dfe"} Dec 05 10:28:46 crc kubenswrapper[4815]: I1205 10:28:46.309189 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcptj" event={"ID":"1da3a037-cf60-474f-be25-62e9601ee7bc","Type":"ContainerStarted","Data":"e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1"} Dec 05 10:28:51 crc kubenswrapper[4815]: I1205 10:28:51.354774 4815 generic.go:334] "Generic (PLEG): container finished" podID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerID="e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1" exitCode=0 Dec 05 10:28:51 crc kubenswrapper[4815]: I1205 10:28:51.354827 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcptj" event={"ID":"1da3a037-cf60-474f-be25-62e9601ee7bc","Type":"ContainerDied","Data":"e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1"} Dec 05 10:28:52 crc kubenswrapper[4815]: I1205 10:28:52.368668 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcptj" event={"ID":"1da3a037-cf60-474f-be25-62e9601ee7bc","Type":"ContainerStarted","Data":"342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c"} Dec 05 10:28:52 crc kubenswrapper[4815]: I1205 10:28:52.400524 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pcptj" podStartSLOduration=4.902733163 podStartE2EDuration="14.400479798s" podCreationTimestamp="2025-12-05 10:28:38 +0000 UTC" firstStartedPulling="2025-12-05 10:28:42.255556184 +0000 UTC m=+4921.134163011" lastFinishedPulling="2025-12-05 10:28:51.753302809 +0000 UTC m=+4930.631909646" observedRunningTime="2025-12-05 10:28:52.397803225 +0000 UTC m=+4931.276410072" watchObservedRunningTime="2025-12-05 10:28:52.400479798 +0000 UTC m=+4931.279086645" Dec 05 10:28:58 crc kubenswrapper[4815]: I1205 10:28:58.429786 4815 generic.go:334] "Generic (PLEG): container finished" podID="7c32a0ff-118b-482f-9e73-15ece3db9f88" containerID="4ff58c55e0ac50c928ea60b59ec9724ba9f03cf007331dc37a3ca34b3973f952" exitCode=0 Dec 05 10:28:58 crc kubenswrapper[4815]: I1205 10:28:58.429882 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" event={"ID":"7c32a0ff-118b-482f-9e73-15ece3db9f88","Type":"ContainerDied","Data":"4ff58c55e0ac50c928ea60b59ec9724ba9f03cf007331dc37a3ca34b3973f952"} Dec 05 10:28:59 crc kubenswrapper[4815]: I1205 10:28:59.194852 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:59 crc kubenswrapper[4815]: I1205 10:28:59.195219 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:28:59 crc kubenswrapper[4815]: I1205 10:28:59.624211 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" Dec 05 10:28:59 crc kubenswrapper[4815]: I1205 10:28:59.665862 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w2mx6/crc-debug-vt2h2"] Dec 05 10:28:59 crc kubenswrapper[4815]: I1205 10:28:59.675294 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w2mx6/crc-debug-vt2h2"] Dec 05 10:28:59 crc kubenswrapper[4815]: I1205 10:28:59.775503 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c32a0ff-118b-482f-9e73-15ece3db9f88-host\") pod \"7c32a0ff-118b-482f-9e73-15ece3db9f88\" (UID: \"7c32a0ff-118b-482f-9e73-15ece3db9f88\") " Dec 05 10:28:59 crc kubenswrapper[4815]: I1205 10:28:59.775598 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dfz62\" (UniqueName: \"kubernetes.io/projected/7c32a0ff-118b-482f-9e73-15ece3db9f88-kube-api-access-dfz62\") pod \"7c32a0ff-118b-482f-9e73-15ece3db9f88\" (UID: \"7c32a0ff-118b-482f-9e73-15ece3db9f88\") " Dec 05 10:28:59 crc kubenswrapper[4815]: I1205 10:28:59.775953 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c32a0ff-118b-482f-9e73-15ece3db9f88-host" (OuterVolumeSpecName: "host") pod "7c32a0ff-118b-482f-9e73-15ece3db9f88" (UID: "7c32a0ff-118b-482f-9e73-15ece3db9f88"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:28:59 crc kubenswrapper[4815]: I1205 10:28:59.776116 4815 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7c32a0ff-118b-482f-9e73-15ece3db9f88-host\") on node \"crc\" DevicePath \"\"" Dec 05 10:28:59 crc kubenswrapper[4815]: I1205 10:28:59.799250 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c32a0ff-118b-482f-9e73-15ece3db9f88-kube-api-access-dfz62" (OuterVolumeSpecName: "kube-api-access-dfz62") pod "7c32a0ff-118b-482f-9e73-15ece3db9f88" (UID: "7c32a0ff-118b-482f-9e73-15ece3db9f88"). InnerVolumeSpecName "kube-api-access-dfz62". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:28:59 crc kubenswrapper[4815]: I1205 10:28:59.878601 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dfz62\" (UniqueName: \"kubernetes.io/projected/7c32a0ff-118b-482f-9e73-15ece3db9f88-kube-api-access-dfz62\") on node \"crc\" DevicePath \"\"" Dec 05 10:29:00 crc kubenswrapper[4815]: I1205 10:29:00.246032 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pcptj" podUID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerName="registry-server" probeResult="failure" output=< Dec 05 10:29:00 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 10:29:00 crc kubenswrapper[4815]: > Dec 05 10:29:00 crc kubenswrapper[4815]: I1205 10:29:00.461871 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab0bfc35b268c388beea787aeb4285666c45568c95a86cafade190bc16654519" Dec 05 10:29:00 crc kubenswrapper[4815]: I1205 10:29:00.461935 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-vt2h2" Dec 05 10:29:00 crc kubenswrapper[4815]: I1205 10:29:00.929835 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w2mx6/crc-debug-mpvpx"] Dec 05 10:29:00 crc kubenswrapper[4815]: E1205 10:29:00.931115 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c32a0ff-118b-482f-9e73-15ece3db9f88" containerName="container-00" Dec 05 10:29:00 crc kubenswrapper[4815]: I1205 10:29:00.931228 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c32a0ff-118b-482f-9e73-15ece3db9f88" containerName="container-00" Dec 05 10:29:00 crc kubenswrapper[4815]: I1205 10:29:00.931632 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c32a0ff-118b-482f-9e73-15ece3db9f88" containerName="container-00" Dec 05 10:29:00 crc kubenswrapper[4815]: I1205 10:29:00.932421 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" Dec 05 10:29:01 crc kubenswrapper[4815]: I1205 10:29:01.104377 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4crs\" (UniqueName: \"kubernetes.io/projected/c09e6412-d53a-498e-90f5-f3d4ba3c2121-kube-api-access-j4crs\") pod \"crc-debug-mpvpx\" (UID: \"c09e6412-d53a-498e-90f5-f3d4ba3c2121\") " pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" Dec 05 10:29:01 crc kubenswrapper[4815]: I1205 10:29:01.104437 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c09e6412-d53a-498e-90f5-f3d4ba3c2121-host\") pod \"crc-debug-mpvpx\" (UID: \"c09e6412-d53a-498e-90f5-f3d4ba3c2121\") " pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" Dec 05 10:29:01 crc kubenswrapper[4815]: I1205 10:29:01.207093 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4crs\" (UniqueName: \"kubernetes.io/projected/c09e6412-d53a-498e-90f5-f3d4ba3c2121-kube-api-access-j4crs\") pod \"crc-debug-mpvpx\" (UID: \"c09e6412-d53a-498e-90f5-f3d4ba3c2121\") " pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" Dec 05 10:29:01 crc kubenswrapper[4815]: I1205 10:29:01.207149 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c09e6412-d53a-498e-90f5-f3d4ba3c2121-host\") pod \"crc-debug-mpvpx\" (UID: \"c09e6412-d53a-498e-90f5-f3d4ba3c2121\") " pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" Dec 05 10:29:01 crc kubenswrapper[4815]: I1205 10:29:01.207290 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c09e6412-d53a-498e-90f5-f3d4ba3c2121-host\") pod \"crc-debug-mpvpx\" (UID: \"c09e6412-d53a-498e-90f5-f3d4ba3c2121\") " pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" Dec 05 10:29:01 crc kubenswrapper[4815]: I1205 10:29:01.440673 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c32a0ff-118b-482f-9e73-15ece3db9f88" path="/var/lib/kubelet/pods/7c32a0ff-118b-482f-9e73-15ece3db9f88/volumes" Dec 05 10:29:02 crc kubenswrapper[4815]: I1205 10:29:02.556466 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4crs\" (UniqueName: \"kubernetes.io/projected/c09e6412-d53a-498e-90f5-f3d4ba3c2121-kube-api-access-j4crs\") pod \"crc-debug-mpvpx\" (UID: \"c09e6412-d53a-498e-90f5-f3d4ba3c2121\") " pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" Dec 05 10:29:02 crc kubenswrapper[4815]: I1205 10:29:02.757982 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" Dec 05 10:29:02 crc kubenswrapper[4815]: W1205 10:29:02.792116 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc09e6412_d53a_498e_90f5_f3d4ba3c2121.slice/crio-d43cbf3eba2add5b75ab235857ece9a01fb3cd2f8fe8a1f41ada304dfd232b42 WatchSource:0}: Error finding container d43cbf3eba2add5b75ab235857ece9a01fb3cd2f8fe8a1f41ada304dfd232b42: Status 404 returned error can't find the container with id d43cbf3eba2add5b75ab235857ece9a01fb3cd2f8fe8a1f41ada304dfd232b42 Dec 05 10:29:03 crc kubenswrapper[4815]: I1205 10:29:03.488391 4815 generic.go:334] "Generic (PLEG): container finished" podID="c09e6412-d53a-498e-90f5-f3d4ba3c2121" containerID="bdcb2f0009bfa9a9bfb619a381278b866f9bc5f8e5de2316901b42ca307b4f1e" exitCode=0 Dec 05 10:29:03 crc kubenswrapper[4815]: I1205 10:29:03.488469 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" event={"ID":"c09e6412-d53a-498e-90f5-f3d4ba3c2121","Type":"ContainerDied","Data":"bdcb2f0009bfa9a9bfb619a381278b866f9bc5f8e5de2316901b42ca307b4f1e"} Dec 05 10:29:03 crc kubenswrapper[4815]: I1205 10:29:03.489114 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" event={"ID":"c09e6412-d53a-498e-90f5-f3d4ba3c2121","Type":"ContainerStarted","Data":"d43cbf3eba2add5b75ab235857ece9a01fb3cd2f8fe8a1f41ada304dfd232b42"} Dec 05 10:29:03 crc kubenswrapper[4815]: I1205 10:29:03.982207 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w2mx6/crc-debug-mpvpx"] Dec 05 10:29:03 crc kubenswrapper[4815]: I1205 10:29:03.993915 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w2mx6/crc-debug-mpvpx"] Dec 05 10:29:04 crc kubenswrapper[4815]: I1205 10:29:04.621034 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" Dec 05 10:29:04 crc kubenswrapper[4815]: I1205 10:29:04.696543 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4crs\" (UniqueName: \"kubernetes.io/projected/c09e6412-d53a-498e-90f5-f3d4ba3c2121-kube-api-access-j4crs\") pod \"c09e6412-d53a-498e-90f5-f3d4ba3c2121\" (UID: \"c09e6412-d53a-498e-90f5-f3d4ba3c2121\") " Dec 05 10:29:04 crc kubenswrapper[4815]: I1205 10:29:04.696884 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c09e6412-d53a-498e-90f5-f3d4ba3c2121-host\") pod \"c09e6412-d53a-498e-90f5-f3d4ba3c2121\" (UID: \"c09e6412-d53a-498e-90f5-f3d4ba3c2121\") " Dec 05 10:29:04 crc kubenswrapper[4815]: I1205 10:29:04.697023 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c09e6412-d53a-498e-90f5-f3d4ba3c2121-host" (OuterVolumeSpecName: "host") pod "c09e6412-d53a-498e-90f5-f3d4ba3c2121" (UID: "c09e6412-d53a-498e-90f5-f3d4ba3c2121"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:29:04 crc kubenswrapper[4815]: I1205 10:29:04.697376 4815 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/c09e6412-d53a-498e-90f5-f3d4ba3c2121-host\") on node \"crc\" DevicePath \"\"" Dec 05 10:29:04 crc kubenswrapper[4815]: I1205 10:29:04.702951 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c09e6412-d53a-498e-90f5-f3d4ba3c2121-kube-api-access-j4crs" (OuterVolumeSpecName: "kube-api-access-j4crs") pod "c09e6412-d53a-498e-90f5-f3d4ba3c2121" (UID: "c09e6412-d53a-498e-90f5-f3d4ba3c2121"). InnerVolumeSpecName "kube-api-access-j4crs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:29:04 crc kubenswrapper[4815]: I1205 10:29:04.799410 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4crs\" (UniqueName: \"kubernetes.io/projected/c09e6412-d53a-498e-90f5-f3d4ba3c2121-kube-api-access-j4crs\") on node \"crc\" DevicePath \"\"" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.154003 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-w2mx6/crc-debug-wlxmt"] Dec 05 10:29:05 crc kubenswrapper[4815]: E1205 10:29:05.154609 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c09e6412-d53a-498e-90f5-f3d4ba3c2121" containerName="container-00" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.154639 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="c09e6412-d53a-498e-90f5-f3d4ba3c2121" containerName="container-00" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.154959 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="c09e6412-d53a-498e-90f5-f3d4ba3c2121" containerName="container-00" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.155967 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.310337 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ba85d2d9-6804-4ce5-84ca-b8157e91d146-host\") pod \"crc-debug-wlxmt\" (UID: \"ba85d2d9-6804-4ce5-84ca-b8157e91d146\") " pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.310523 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6c77j\" (UniqueName: \"kubernetes.io/projected/ba85d2d9-6804-4ce5-84ca-b8157e91d146-kube-api-access-6c77j\") pod \"crc-debug-wlxmt\" (UID: \"ba85d2d9-6804-4ce5-84ca-b8157e91d146\") " pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.412724 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ba85d2d9-6804-4ce5-84ca-b8157e91d146-host\") pod \"crc-debug-wlxmt\" (UID: \"ba85d2d9-6804-4ce5-84ca-b8157e91d146\") " pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.412861 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6c77j\" (UniqueName: \"kubernetes.io/projected/ba85d2d9-6804-4ce5-84ca-b8157e91d146-kube-api-access-6c77j\") pod \"crc-debug-wlxmt\" (UID: \"ba85d2d9-6804-4ce5-84ca-b8157e91d146\") " pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.412882 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ba85d2d9-6804-4ce5-84ca-b8157e91d146-host\") pod \"crc-debug-wlxmt\" (UID: \"ba85d2d9-6804-4ce5-84ca-b8157e91d146\") " pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.430654 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c09e6412-d53a-498e-90f5-f3d4ba3c2121" path="/var/lib/kubelet/pods/c09e6412-d53a-498e-90f5-f3d4ba3c2121/volumes" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.447714 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6c77j\" (UniqueName: \"kubernetes.io/projected/ba85d2d9-6804-4ce5-84ca-b8157e91d146-kube-api-access-6c77j\") pod \"crc-debug-wlxmt\" (UID: \"ba85d2d9-6804-4ce5-84ca-b8157e91d146\") " pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.472870 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" Dec 05 10:29:05 crc kubenswrapper[4815]: W1205 10:29:05.506946 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba85d2d9_6804_4ce5_84ca_b8157e91d146.slice/crio-2ea8f2cd93bcbd052825a2de0584261b86a0d3e119fde970c90389efce892cdd WatchSource:0}: Error finding container 2ea8f2cd93bcbd052825a2de0584261b86a0d3e119fde970c90389efce892cdd: Status 404 returned error can't find the container with id 2ea8f2cd93bcbd052825a2de0584261b86a0d3e119fde970c90389efce892cdd Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.507711 4815 scope.go:117] "RemoveContainer" containerID="bdcb2f0009bfa9a9bfb619a381278b866f9bc5f8e5de2316901b42ca307b4f1e" Dec 05 10:29:05 crc kubenswrapper[4815]: I1205 10:29:05.507750 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-mpvpx" Dec 05 10:29:06 crc kubenswrapper[4815]: I1205 10:29:06.520023 4815 generic.go:334] "Generic (PLEG): container finished" podID="ba85d2d9-6804-4ce5-84ca-b8157e91d146" containerID="69d978d6232ec92f4755f3b8621e16f92a68c751505486e64ca7d2b40fa2288c" exitCode=0 Dec 05 10:29:06 crc kubenswrapper[4815]: I1205 10:29:06.520091 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" event={"ID":"ba85d2d9-6804-4ce5-84ca-b8157e91d146","Type":"ContainerDied","Data":"69d978d6232ec92f4755f3b8621e16f92a68c751505486e64ca7d2b40fa2288c"} Dec 05 10:29:06 crc kubenswrapper[4815]: I1205 10:29:06.520322 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" event={"ID":"ba85d2d9-6804-4ce5-84ca-b8157e91d146","Type":"ContainerStarted","Data":"2ea8f2cd93bcbd052825a2de0584261b86a0d3e119fde970c90389efce892cdd"} Dec 05 10:29:06 crc kubenswrapper[4815]: I1205 10:29:06.563964 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w2mx6/crc-debug-wlxmt"] Dec 05 10:29:06 crc kubenswrapper[4815]: I1205 10:29:06.578320 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w2mx6/crc-debug-wlxmt"] Dec 05 10:29:07 crc kubenswrapper[4815]: I1205 10:29:07.838853 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="ae4a421a-317a-4a4b-bcb1-2717aba4d25f" containerName="galera" probeResult="failure" output="command timed out" Dec 05 10:29:07 crc kubenswrapper[4815]: I1205 10:29:07.839073 4815 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="ae4a421a-317a-4a4b-bcb1-2717aba4d25f" containerName="galera" probeResult="failure" output="command timed out" Dec 05 10:29:07 crc kubenswrapper[4815]: I1205 10:29:07.950915 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" Dec 05 10:29:07 crc kubenswrapper[4815]: I1205 10:29:07.981136 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ba85d2d9-6804-4ce5-84ca-b8157e91d146-host\") pod \"ba85d2d9-6804-4ce5-84ca-b8157e91d146\" (UID: \"ba85d2d9-6804-4ce5-84ca-b8157e91d146\") " Dec 05 10:29:07 crc kubenswrapper[4815]: I1205 10:29:07.981229 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ba85d2d9-6804-4ce5-84ca-b8157e91d146-host" (OuterVolumeSpecName: "host") pod "ba85d2d9-6804-4ce5-84ca-b8157e91d146" (UID: "ba85d2d9-6804-4ce5-84ca-b8157e91d146"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:29:07 crc kubenswrapper[4815]: I1205 10:29:07.981575 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6c77j\" (UniqueName: \"kubernetes.io/projected/ba85d2d9-6804-4ce5-84ca-b8157e91d146-kube-api-access-6c77j\") pod \"ba85d2d9-6804-4ce5-84ca-b8157e91d146\" (UID: \"ba85d2d9-6804-4ce5-84ca-b8157e91d146\") " Dec 05 10:29:07 crc kubenswrapper[4815]: I1205 10:29:07.982558 4815 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/ba85d2d9-6804-4ce5-84ca-b8157e91d146-host\") on node \"crc\" DevicePath \"\"" Dec 05 10:29:07 crc kubenswrapper[4815]: I1205 10:29:07.988746 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba85d2d9-6804-4ce5-84ca-b8157e91d146-kube-api-access-6c77j" (OuterVolumeSpecName: "kube-api-access-6c77j") pod "ba85d2d9-6804-4ce5-84ca-b8157e91d146" (UID: "ba85d2d9-6804-4ce5-84ca-b8157e91d146"). InnerVolumeSpecName "kube-api-access-6c77j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:29:08 crc kubenswrapper[4815]: I1205 10:29:08.084532 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6c77j\" (UniqueName: \"kubernetes.io/projected/ba85d2d9-6804-4ce5-84ca-b8157e91d146-kube-api-access-6c77j\") on node \"crc\" DevicePath \"\"" Dec 05 10:29:08 crc kubenswrapper[4815]: I1205 10:29:08.541379 4815 scope.go:117] "RemoveContainer" containerID="69d978d6232ec92f4755f3b8621e16f92a68c751505486e64ca7d2b40fa2288c" Dec 05 10:29:08 crc kubenswrapper[4815]: I1205 10:29:08.541550 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/crc-debug-wlxmt" Dec 05 10:29:09 crc kubenswrapper[4815]: I1205 10:29:09.266711 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:29:09 crc kubenswrapper[4815]: I1205 10:29:09.440281 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba85d2d9-6804-4ce5-84ca-b8157e91d146" path="/var/lib/kubelet/pods/ba85d2d9-6804-4ce5-84ca-b8157e91d146/volumes" Dec 05 10:29:09 crc kubenswrapper[4815]: I1205 10:29:09.761340 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:29:10 crc kubenswrapper[4815]: I1205 10:29:10.043387 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pcptj"] Dec 05 10:29:10 crc kubenswrapper[4815]: I1205 10:29:10.563041 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pcptj" podUID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerName="registry-server" containerID="cri-o://342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c" gracePeriod=2 Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.024739 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.115897 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txsr6\" (UniqueName: \"kubernetes.io/projected/1da3a037-cf60-474f-be25-62e9601ee7bc-kube-api-access-txsr6\") pod \"1da3a037-cf60-474f-be25-62e9601ee7bc\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.116276 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-utilities\") pod \"1da3a037-cf60-474f-be25-62e9601ee7bc\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.116561 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-catalog-content\") pod \"1da3a037-cf60-474f-be25-62e9601ee7bc\" (UID: \"1da3a037-cf60-474f-be25-62e9601ee7bc\") " Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.117064 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-utilities" (OuterVolumeSpecName: "utilities") pod "1da3a037-cf60-474f-be25-62e9601ee7bc" (UID: "1da3a037-cf60-474f-be25-62e9601ee7bc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.117233 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.127035 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1da3a037-cf60-474f-be25-62e9601ee7bc-kube-api-access-txsr6" (OuterVolumeSpecName: "kube-api-access-txsr6") pod "1da3a037-cf60-474f-be25-62e9601ee7bc" (UID: "1da3a037-cf60-474f-be25-62e9601ee7bc"). InnerVolumeSpecName "kube-api-access-txsr6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.219540 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txsr6\" (UniqueName: \"kubernetes.io/projected/1da3a037-cf60-474f-be25-62e9601ee7bc-kube-api-access-txsr6\") on node \"crc\" DevicePath \"\"" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.270212 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1da3a037-cf60-474f-be25-62e9601ee7bc" (UID: "1da3a037-cf60-474f-be25-62e9601ee7bc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.321522 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1da3a037-cf60-474f-be25-62e9601ee7bc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.575209 4815 generic.go:334] "Generic (PLEG): container finished" podID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerID="342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c" exitCode=0 Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.575254 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcptj" event={"ID":"1da3a037-cf60-474f-be25-62e9601ee7bc","Type":"ContainerDied","Data":"342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c"} Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.575283 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pcptj" event={"ID":"1da3a037-cf60-474f-be25-62e9601ee7bc","Type":"ContainerDied","Data":"b014e9932bae82d1ee5a5dd7d04fcb28f7a0bd4fbe405bdbd309889e0e42ffef"} Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.575292 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pcptj" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.575302 4815 scope.go:117] "RemoveContainer" containerID="342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.608166 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pcptj"] Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.618402 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pcptj"] Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.619244 4815 scope.go:117] "RemoveContainer" containerID="e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.643353 4815 scope.go:117] "RemoveContainer" containerID="fed48841608062541922634baff7c538ecef5ef6786b478fd45588f0a2296dfe" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.704341 4815 scope.go:117] "RemoveContainer" containerID="342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c" Dec 05 10:29:11 crc kubenswrapper[4815]: E1205 10:29:11.704847 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c\": container with ID starting with 342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c not found: ID does not exist" containerID="342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.704907 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c"} err="failed to get container status \"342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c\": rpc error: code = NotFound desc = could not find container \"342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c\": container with ID starting with 342006033fdb14ba0ae1fae5804868f2c028f0193166d599b13d7eee4617352c not found: ID does not exist" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.704946 4815 scope.go:117] "RemoveContainer" containerID="e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1" Dec 05 10:29:11 crc kubenswrapper[4815]: E1205 10:29:11.705257 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1\": container with ID starting with e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1 not found: ID does not exist" containerID="e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.705288 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1"} err="failed to get container status \"e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1\": rpc error: code = NotFound desc = could not find container \"e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1\": container with ID starting with e197600d340a70fc60c5015148af14a2784bfa76044fcf38c0f2b42eae7791f1 not found: ID does not exist" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.705311 4815 scope.go:117] "RemoveContainer" containerID="fed48841608062541922634baff7c538ecef5ef6786b478fd45588f0a2296dfe" Dec 05 10:29:11 crc kubenswrapper[4815]: E1205 10:29:11.705569 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fed48841608062541922634baff7c538ecef5ef6786b478fd45588f0a2296dfe\": container with ID starting with fed48841608062541922634baff7c538ecef5ef6786b478fd45588f0a2296dfe not found: ID does not exist" containerID="fed48841608062541922634baff7c538ecef5ef6786b478fd45588f0a2296dfe" Dec 05 10:29:11 crc kubenswrapper[4815]: I1205 10:29:11.705590 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fed48841608062541922634baff7c538ecef5ef6786b478fd45588f0a2296dfe"} err="failed to get container status \"fed48841608062541922634baff7c538ecef5ef6786b478fd45588f0a2296dfe\": rpc error: code = NotFound desc = could not find container \"fed48841608062541922634baff7c538ecef5ef6786b478fd45588f0a2296dfe\": container with ID starting with fed48841608062541922634baff7c538ecef5ef6786b478fd45588f0a2296dfe not found: ID does not exist" Dec 05 10:29:13 crc kubenswrapper[4815]: I1205 10:29:13.432267 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1da3a037-cf60-474f-be25-62e9601ee7bc" path="/var/lib/kubelet/pods/1da3a037-cf60-474f-be25-62e9601ee7bc/volumes" Dec 05 10:29:35 crc kubenswrapper[4815]: I1205 10:29:35.751239 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-588897d4d8-8nmk9_d7266a34-7fa1-442b-9e83-e66665a5cd12/barbican-api/0.log" Dec 05 10:29:35 crc kubenswrapper[4815]: I1205 10:29:35.912762 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-588897d4d8-8nmk9_d7266a34-7fa1-442b-9e83-e66665a5cd12/barbican-api-log/0.log" Dec 05 10:29:36 crc kubenswrapper[4815]: I1205 10:29:36.011898 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-596686cb9d-rwjpw_3f638151-76e9-429e-b283-d741ad857276/barbican-keystone-listener-log/0.log" Dec 05 10:29:36 crc kubenswrapper[4815]: I1205 10:29:36.016658 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-596686cb9d-rwjpw_3f638151-76e9-429e-b283-d741ad857276/barbican-keystone-listener/0.log" Dec 05 10:29:36 crc kubenswrapper[4815]: I1205 10:29:36.226049 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57db4d6df7-gdwss_92f1b7e3-1999-448d-90c1-9845b162701b/barbican-worker-log/0.log" Dec 05 10:29:36 crc kubenswrapper[4815]: I1205 10:29:36.289500 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57db4d6df7-gdwss_92f1b7e3-1999-448d-90c1-9845b162701b/barbican-worker/0.log" Dec 05 10:29:36 crc kubenswrapper[4815]: I1205 10:29:36.698648 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7_296de34a-3504-4fb8-8a20-02333f447f4c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:36 crc kubenswrapper[4815]: I1205 10:29:36.788752 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_078b665f-5e54-4d83-966b-3684d06ea320/ceilometer-central-agent/0.log" Dec 05 10:29:36 crc kubenswrapper[4815]: I1205 10:29:36.840814 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_078b665f-5e54-4d83-966b-3684d06ea320/ceilometer-notification-agent/0.log" Dec 05 10:29:36 crc kubenswrapper[4815]: I1205 10:29:36.918730 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_078b665f-5e54-4d83-966b-3684d06ea320/proxy-httpd/0.log" Dec 05 10:29:37 crc kubenswrapper[4815]: I1205 10:29:37.028023 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_078b665f-5e54-4d83-966b-3684d06ea320/sg-core/0.log" Dec 05 10:29:37 crc kubenswrapper[4815]: I1205 10:29:37.077194 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8_dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:37 crc kubenswrapper[4815]: I1205 10:29:37.335781 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r_18189644-5dd2-4329-a10b-70d62c5e0fd0/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:37 crc kubenswrapper[4815]: I1205 10:29:37.400252 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05/cinder-api/0.log" Dec 05 10:29:37 crc kubenswrapper[4815]: I1205 10:29:37.486368 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05/cinder-api-log/0.log" Dec 05 10:29:37 crc kubenswrapper[4815]: I1205 10:29:37.648153 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_6f53fc9d-0e41-4083-855a-1624fb520e84/probe/0.log" Dec 05 10:29:37 crc kubenswrapper[4815]: I1205 10:29:37.777674 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_6f53fc9d-0e41-4083-855a-1624fb520e84/cinder-backup/0.log" Dec 05 10:29:37 crc kubenswrapper[4815]: I1205 10:29:37.923437 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9a6d4017-ba13-47e9-b0e2-069a6d6afa21/cinder-scheduler/0.log" Dec 05 10:29:37 crc kubenswrapper[4815]: I1205 10:29:37.946095 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9a6d4017-ba13-47e9-b0e2-069a6d6afa21/probe/0.log" Dec 05 10:29:38 crc kubenswrapper[4815]: I1205 10:29:38.045989 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_15d9387d-aef1-48fe-8fb1-cb9d3dec22ea/cinder-volume/0.log" Dec 05 10:29:38 crc kubenswrapper[4815]: I1205 10:29:38.204294 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_15d9387d-aef1-48fe-8fb1-cb9d3dec22ea/probe/0.log" Dec 05 10:29:38 crc kubenswrapper[4815]: I1205 10:29:38.273470 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-699sn_06fceca7-a988-4d36-8a56-fbe8fb96a28b/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:38 crc kubenswrapper[4815]: I1205 10:29:38.805188 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh_4ab36e54-e39b-4b61-8609-52004d810c7a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:38 crc kubenswrapper[4815]: I1205 10:29:38.886978 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-94c764695-ml6b2_310bad94-91d6-4420-9d0c-8ed3af8b72c5/init/0.log" Dec 05 10:29:39 crc kubenswrapper[4815]: I1205 10:29:39.079029 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-94c764695-ml6b2_310bad94-91d6-4420-9d0c-8ed3af8b72c5/init/0.log" Dec 05 10:29:39 crc kubenswrapper[4815]: I1205 10:29:39.235794 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b/glance-httpd/0.log" Dec 05 10:29:39 crc kubenswrapper[4815]: I1205 10:29:39.323625 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-94c764695-ml6b2_310bad94-91d6-4420-9d0c-8ed3af8b72c5/dnsmasq-dns/0.log" Dec 05 10:29:39 crc kubenswrapper[4815]: I1205 10:29:39.398400 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b/glance-log/0.log" Dec 05 10:29:39 crc kubenswrapper[4815]: I1205 10:29:39.455972 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7330ea7b-3369-4ce7-95b6-fa415d3255f8/glance-httpd/0.log" Dec 05 10:29:39 crc kubenswrapper[4815]: I1205 10:29:39.553720 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7330ea7b-3369-4ce7-95b6-fa415d3255f8/glance-log/0.log" Dec 05 10:29:39 crc kubenswrapper[4815]: I1205 10:29:39.887240 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-74b8f449c4-9gwqr_e9604337-767f-492a-aa51-f31578f44ece/horizon/0.log" Dec 05 10:29:39 crc kubenswrapper[4815]: I1205 10:29:39.970814 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-74b8f449c4-9gwqr_e9604337-767f-492a-aa51-f31578f44ece/horizon-log/0.log" Dec 05 10:29:40 crc kubenswrapper[4815]: I1205 10:29:40.073997 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9_abe1e909-bfe1-4bab-9b1e-5ed3316288b6/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:41 crc kubenswrapper[4815]: I1205 10:29:41.419952 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-cx6hg_711134c0-16f2-4ba4-8120-1c39b9d833b2/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:41 crc kubenswrapper[4815]: I1205 10:29:41.545563 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29415481-7x47p_fba6bed7-9ef2-408d-ac41-04bff23ec468/keystone-cron/0.log" Dec 05 10:29:41 crc kubenswrapper[4815]: I1205 10:29:41.696689 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6f56547bd4-2dxhd_7518e451-ec15-4df9-8fa0-842ca6094b36/keystone-api/0.log" Dec 05 10:29:41 crc kubenswrapper[4815]: I1205 10:29:41.699046 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b7f328c9-41c2-4d4d-9544-c22bcc026c33/kube-state-metrics/0.log" Dec 05 10:29:41 crc kubenswrapper[4815]: I1205 10:29:41.816905 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-kknjs_72f92018-4b16-474a-b5de-8fd124dd857b/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:42 crc kubenswrapper[4815]: I1205 10:29:42.008348 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_de4b5d94-2c11-4552-8230-08b776caa2a8/manila-api-log/0.log" Dec 05 10:29:42 crc kubenswrapper[4815]: I1205 10:29:42.112855 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_de4b5d94-2c11-4552-8230-08b776caa2a8/manila-api/0.log" Dec 05 10:29:42 crc kubenswrapper[4815]: I1205 10:29:42.169070 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_bed98f1a-2e43-47cc-82af-ca5c1a833995/manila-scheduler/0.log" Dec 05 10:29:42 crc kubenswrapper[4815]: I1205 10:29:42.244252 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_bed98f1a-2e43-47cc-82af-ca5c1a833995/probe/0.log" Dec 05 10:29:42 crc kubenswrapper[4815]: I1205 10:29:42.406937 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7d69d7b9-98dc-4dff-96bc-e562b8bc4b98/manila-share/0.log" Dec 05 10:29:42 crc kubenswrapper[4815]: I1205 10:29:42.420236 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7d69d7b9-98dc-4dff-96bc-e562b8bc4b98/probe/0.log" Dec 05 10:29:42 crc kubenswrapper[4815]: I1205 10:29:42.817396 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54c4865f77-tppk8_fceefc10-15e5-4b68-9bb6-e786dde63941/neutron-httpd/0.log" Dec 05 10:29:42 crc kubenswrapper[4815]: I1205 10:29:42.892188 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54c4865f77-tppk8_fceefc10-15e5-4b68-9bb6-e786dde63941/neutron-api/0.log" Dec 05 10:29:42 crc kubenswrapper[4815]: I1205 10:29:42.939714 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n_76992009-c2d1-4e3b-be35-70fc39e9f999/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:43 crc kubenswrapper[4815]: I1205 10:29:43.419115 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_7d358cc7-a297-4407-85db-ca619d6dccd4/nova-cell0-conductor-conductor/0.log" Dec 05 10:29:43 crc kubenswrapper[4815]: I1205 10:29:43.571472 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_29c55b4a-b8d1-4604-ad9a-de77d453795f/nova-api-log/0.log" Dec 05 10:29:43 crc kubenswrapper[4815]: I1205 10:29:43.853609 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_62b5fcba-8ff7-4438-8b1c-960c0996a271/nova-cell1-conductor-conductor/0.log" Dec 05 10:29:43 crc kubenswrapper[4815]: I1205 10:29:43.948874 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_29c55b4a-b8d1-4604-ad9a-de77d453795f/nova-api-api/0.log" Dec 05 10:29:44 crc kubenswrapper[4815]: I1205 10:29:44.000018 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_eea3bd17-9d01-4cf4-bdb2-4f5a48a25581/nova-cell1-novncproxy-novncproxy/0.log" Dec 05 10:29:44 crc kubenswrapper[4815]: I1205 10:29:44.295001 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl_e6418e66-6e27-4292-88c6-fca958ba3665/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:44 crc kubenswrapper[4815]: I1205 10:29:44.396710 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a7daef92-258a-4898-890a-44de7ad290b3/nova-metadata-log/0.log" Dec 05 10:29:44 crc kubenswrapper[4815]: I1205 10:29:44.853249 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1/mysql-bootstrap/0.log" Dec 05 10:29:44 crc kubenswrapper[4815]: I1205 10:29:44.869711 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_89bbf450-abe1-4daa-8060-40077543ac23/nova-scheduler-scheduler/0.log" Dec 05 10:29:45 crc kubenswrapper[4815]: I1205 10:29:45.032309 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1/mysql-bootstrap/0.log" Dec 05 10:29:45 crc kubenswrapper[4815]: I1205 10:29:45.135588 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1/galera/0.log" Dec 05 10:29:45 crc kubenswrapper[4815]: I1205 10:29:45.413108 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae4a421a-317a-4a4b-bcb1-2717aba4d25f/mysql-bootstrap/0.log" Dec 05 10:29:45 crc kubenswrapper[4815]: I1205 10:29:45.664173 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae4a421a-317a-4a4b-bcb1-2717aba4d25f/galera/0.log" Dec 05 10:29:45 crc kubenswrapper[4815]: I1205 10:29:45.703467 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae4a421a-317a-4a4b-bcb1-2717aba4d25f/mysql-bootstrap/0.log" Dec 05 10:29:45 crc kubenswrapper[4815]: I1205 10:29:45.889761 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_04cdd6cd-6769-4296-a45d-fffd016fdfd6/openstackclient/0.log" Dec 05 10:29:46 crc kubenswrapper[4815]: I1205 10:29:46.073563 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-6q6fd_0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0/openstack-network-exporter/0.log" Dec 05 10:29:46 crc kubenswrapper[4815]: I1205 10:29:46.275723 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zlv59_6e811382-f977-41d3-aa18-65fe948261dd/ovsdb-server-init/0.log" Dec 05 10:29:46 crc kubenswrapper[4815]: I1205 10:29:46.472841 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a7daef92-258a-4898-890a-44de7ad290b3/nova-metadata-metadata/0.log" Dec 05 10:29:46 crc kubenswrapper[4815]: I1205 10:29:46.526907 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zlv59_6e811382-f977-41d3-aa18-65fe948261dd/ovs-vswitchd/0.log" Dec 05 10:29:46 crc kubenswrapper[4815]: I1205 10:29:46.590113 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zlv59_6e811382-f977-41d3-aa18-65fe948261dd/ovsdb-server-init/0.log" Dec 05 10:29:46 crc kubenswrapper[4815]: I1205 10:29:46.695835 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zlv59_6e811382-f977-41d3-aa18-65fe948261dd/ovsdb-server/0.log" Dec 05 10:29:46 crc kubenswrapper[4815]: I1205 10:29:46.912718 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-v5mcb_0e0fc690-fb2a-4381-93df-1adcb21c0c7d/ovn-controller/0.log" Dec 05 10:29:47 crc kubenswrapper[4815]: I1205 10:29:47.131056 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-kv4qr_b3cdccbb-a413-4128-8d79-34ac79a10a45/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:47 crc kubenswrapper[4815]: I1205 10:29:47.307433 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c3e63e92-1982-4055-9df7-db07096a570e/openstack-network-exporter/0.log" Dec 05 10:29:47 crc kubenswrapper[4815]: I1205 10:29:47.349046 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c3e63e92-1982-4055-9df7-db07096a570e/ovn-northd/0.log" Dec 05 10:29:47 crc kubenswrapper[4815]: I1205 10:29:47.643930 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a065bb70-93f8-4e26-8ce0-6f92b7d7cba6/openstack-network-exporter/0.log" Dec 05 10:29:47 crc kubenswrapper[4815]: I1205 10:29:47.936590 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_82494398-26f8-416f-88cd-65a7365cd2c2/openstack-network-exporter/0.log" Dec 05 10:29:47 crc kubenswrapper[4815]: I1205 10:29:47.997401 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_82494398-26f8-416f-88cd-65a7365cd2c2/ovsdbserver-sb/0.log" Dec 05 10:29:48 crc kubenswrapper[4815]: I1205 10:29:48.001289 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a065bb70-93f8-4e26-8ce0-6f92b7d7cba6/ovsdbserver-nb/0.log" Dec 05 10:29:48 crc kubenswrapper[4815]: I1205 10:29:48.285040 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-549777fc4b-mjkbk_75817271-e950-4196-bc2c-b32ac4dad684/placement-api/0.log" Dec 05 10:29:48 crc kubenswrapper[4815]: I1205 10:29:48.437945 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_db8bc583-6bcf-40c7-875b-d5c9544af543/setup-container/0.log" Dec 05 10:29:48 crc kubenswrapper[4815]: I1205 10:29:48.493691 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-549777fc4b-mjkbk_75817271-e950-4196-bc2c-b32ac4dad684/placement-log/0.log" Dec 05 10:29:49 crc kubenswrapper[4815]: I1205 10:29:49.257368 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_db8bc583-6bcf-40c7-875b-d5c9544af543/rabbitmq/0.log" Dec 05 10:29:49 crc kubenswrapper[4815]: I1205 10:29:49.305949 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_db8bc583-6bcf-40c7-875b-d5c9544af543/setup-container/0.log" Dec 05 10:29:49 crc kubenswrapper[4815]: I1205 10:29:49.321000 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d0d7070e-5d74-4d40-9fe9-b42c60585002/setup-container/0.log" Dec 05 10:29:49 crc kubenswrapper[4815]: I1205 10:29:49.662505 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d0d7070e-5d74-4d40-9fe9-b42c60585002/rabbitmq/0.log" Dec 05 10:29:49 crc kubenswrapper[4815]: I1205 10:29:49.697667 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d0d7070e-5d74-4d40-9fe9-b42c60585002/setup-container/0.log" Dec 05 10:29:49 crc kubenswrapper[4815]: I1205 10:29:49.801976 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz_e775d3b4-8e2f-4255-8175-9664129b7998/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:50 crc kubenswrapper[4815]: I1205 10:29:50.098446 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s_0e0848e7-a6e0-4672-bb56-2f40193be881/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:50 crc kubenswrapper[4815]: I1205 10:29:50.219475 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-46rkt_7251f2d6-ee27-4b5d-9d1a-6aeab813638b/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:29:50 crc kubenswrapper[4815]: I1205 10:29:50.440516 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-w8b66_ba6c4e35-e275-4cd3-8f30-0102966dcc1b/ssh-known-hosts-edpm-deployment/0.log" Dec 05 10:29:51 crc kubenswrapper[4815]: I1205 10:29:51.132417 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_b0ec9924-8223-4a07-806b-0ee6451a7021/tempest-tests-tempest-tests-runner/0.log" Dec 05 10:29:51 crc kubenswrapper[4815]: I1205 10:29:51.191479 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_6fb2333c-6791-4164-a041-fec86a88cc06/test-operator-logs-container/0.log" Dec 05 10:29:51 crc kubenswrapper[4815]: I1205 10:29:51.523845 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-f54t2_abc71b71-d4f3-4c99-8a40-8969b4371f39/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.593277 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7"] Dec 05 10:30:00 crc kubenswrapper[4815]: E1205 10:30:00.594173 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerName="extract-content" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.594194 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerName="extract-content" Dec 05 10:30:00 crc kubenswrapper[4815]: E1205 10:30:00.594210 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerName="extract-utilities" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.594216 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerName="extract-utilities" Dec 05 10:30:00 crc kubenswrapper[4815]: E1205 10:30:00.594241 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerName="registry-server" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.594247 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerName="registry-server" Dec 05 10:30:00 crc kubenswrapper[4815]: E1205 10:30:00.594267 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba85d2d9-6804-4ce5-84ca-b8157e91d146" containerName="container-00" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.594273 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba85d2d9-6804-4ce5-84ca-b8157e91d146" containerName="container-00" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.594520 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1da3a037-cf60-474f-be25-62e9601ee7bc" containerName="registry-server" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.594549 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba85d2d9-6804-4ce5-84ca-b8157e91d146" containerName="container-00" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.595270 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.605422 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7"] Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.606441 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.615384 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.694778 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzxxl\" (UniqueName: \"kubernetes.io/projected/1660774c-b853-4b8e-8e44-e581e3623a67-kube-api-access-vzxxl\") pod \"collect-profiles-29415510-bfjp7\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.695202 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1660774c-b853-4b8e-8e44-e581e3623a67-secret-volume\") pod \"collect-profiles-29415510-bfjp7\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.695346 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1660774c-b853-4b8e-8e44-e581e3623a67-config-volume\") pod \"collect-profiles-29415510-bfjp7\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.799374 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1660774c-b853-4b8e-8e44-e581e3623a67-config-volume\") pod \"collect-profiles-29415510-bfjp7\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.799696 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzxxl\" (UniqueName: \"kubernetes.io/projected/1660774c-b853-4b8e-8e44-e581e3623a67-kube-api-access-vzxxl\") pod \"collect-profiles-29415510-bfjp7\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.799801 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1660774c-b853-4b8e-8e44-e581e3623a67-secret-volume\") pod \"collect-profiles-29415510-bfjp7\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.802335 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1660774c-b853-4b8e-8e44-e581e3623a67-config-volume\") pod \"collect-profiles-29415510-bfjp7\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.909309 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1660774c-b853-4b8e-8e44-e581e3623a67-secret-volume\") pod \"collect-profiles-29415510-bfjp7\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:00 crc kubenswrapper[4815]: I1205 10:30:00.942300 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzxxl\" (UniqueName: \"kubernetes.io/projected/1660774c-b853-4b8e-8e44-e581e3623a67-kube-api-access-vzxxl\") pod \"collect-profiles-29415510-bfjp7\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:01 crc kubenswrapper[4815]: I1205 10:30:01.223748 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:01 crc kubenswrapper[4815]: I1205 10:30:01.794322 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7"] Dec 05 10:30:02 crc kubenswrapper[4815]: I1205 10:30:02.222469 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" event={"ID":"1660774c-b853-4b8e-8e44-e581e3623a67","Type":"ContainerStarted","Data":"e5970d6c0030811e7345234f8846873ab61e86decb2ffe2b3c53374323380a30"} Dec 05 10:30:02 crc kubenswrapper[4815]: I1205 10:30:02.222548 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" event={"ID":"1660774c-b853-4b8e-8e44-e581e3623a67","Type":"ContainerStarted","Data":"8bda031ca66d0a83644baadba3e4568fb868a7e3f3a721292ad4110aa5287b46"} Dec 05 10:30:02 crc kubenswrapper[4815]: I1205 10:30:02.245865 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" podStartSLOduration=2.245848735 podStartE2EDuration="2.245848735s" podCreationTimestamp="2025-12-05 10:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:30:02.241236249 +0000 UTC m=+5001.119843086" watchObservedRunningTime="2025-12-05 10:30:02.245848735 +0000 UTC m=+5001.124455572" Dec 05 10:30:03 crc kubenswrapper[4815]: I1205 10:30:03.240179 4815 generic.go:334] "Generic (PLEG): container finished" podID="1660774c-b853-4b8e-8e44-e581e3623a67" containerID="e5970d6c0030811e7345234f8846873ab61e86decb2ffe2b3c53374323380a30" exitCode=0 Dec 05 10:30:03 crc kubenswrapper[4815]: I1205 10:30:03.240534 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" event={"ID":"1660774c-b853-4b8e-8e44-e581e3623a67","Type":"ContainerDied","Data":"e5970d6c0030811e7345234f8846873ab61e86decb2ffe2b3c53374323380a30"} Dec 05 10:30:04 crc kubenswrapper[4815]: I1205 10:30:04.982381 4815 trace.go:236] Trace[1356635527]: "Calculate volume metrics of catalog-content for pod openshift-marketplace/certified-operators-l4v4h" (05-Dec-2025 10:30:03.941) (total time: 1032ms): Dec 05 10:30:04 crc kubenswrapper[4815]: Trace[1356635527]: [1.032295423s] [1.032295423s] END Dec 05 10:30:05 crc kubenswrapper[4815]: I1205 10:30:05.364112 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:05 crc kubenswrapper[4815]: I1205 10:30:05.515201 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1660774c-b853-4b8e-8e44-e581e3623a67-config-volume\") pod \"1660774c-b853-4b8e-8e44-e581e3623a67\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " Dec 05 10:30:05 crc kubenswrapper[4815]: I1205 10:30:05.516587 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1660774c-b853-4b8e-8e44-e581e3623a67-config-volume" (OuterVolumeSpecName: "config-volume") pod "1660774c-b853-4b8e-8e44-e581e3623a67" (UID: "1660774c-b853-4b8e-8e44-e581e3623a67"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:30:05 crc kubenswrapper[4815]: I1205 10:30:05.519147 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzxxl\" (UniqueName: \"kubernetes.io/projected/1660774c-b853-4b8e-8e44-e581e3623a67-kube-api-access-vzxxl\") pod \"1660774c-b853-4b8e-8e44-e581e3623a67\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " Dec 05 10:30:05 crc kubenswrapper[4815]: I1205 10:30:05.519580 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1660774c-b853-4b8e-8e44-e581e3623a67-secret-volume\") pod \"1660774c-b853-4b8e-8e44-e581e3623a67\" (UID: \"1660774c-b853-4b8e-8e44-e581e3623a67\") " Dec 05 10:30:05 crc kubenswrapper[4815]: I1205 10:30:05.520403 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/1660774c-b853-4b8e-8e44-e581e3623a67-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:05 crc kubenswrapper[4815]: I1205 10:30:05.538005 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1660774c-b853-4b8e-8e44-e581e3623a67-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "1660774c-b853-4b8e-8e44-e581e3623a67" (UID: "1660774c-b853-4b8e-8e44-e581e3623a67"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:30:05 crc kubenswrapper[4815]: I1205 10:30:05.545218 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1660774c-b853-4b8e-8e44-e581e3623a67-kube-api-access-vzxxl" (OuterVolumeSpecName: "kube-api-access-vzxxl") pod "1660774c-b853-4b8e-8e44-e581e3623a67" (UID: "1660774c-b853-4b8e-8e44-e581e3623a67"). InnerVolumeSpecName "kube-api-access-vzxxl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:30:05 crc kubenswrapper[4815]: I1205 10:30:05.622882 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzxxl\" (UniqueName: \"kubernetes.io/projected/1660774c-b853-4b8e-8e44-e581e3623a67-kube-api-access-vzxxl\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:05 crc kubenswrapper[4815]: I1205 10:30:05.623117 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/1660774c-b853-4b8e-8e44-e581e3623a67-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:06 crc kubenswrapper[4815]: I1205 10:30:06.266306 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" event={"ID":"1660774c-b853-4b8e-8e44-e581e3623a67","Type":"ContainerDied","Data":"8bda031ca66d0a83644baadba3e4568fb868a7e3f3a721292ad4110aa5287b46"} Dec 05 10:30:06 crc kubenswrapper[4815]: I1205 10:30:06.266363 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bda031ca66d0a83644baadba3e4568fb868a7e3f3a721292ad4110aa5287b46" Dec 05 10:30:06 crc kubenswrapper[4815]: I1205 10:30:06.266427 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415510-bfjp7" Dec 05 10:30:06 crc kubenswrapper[4815]: I1205 10:30:06.466409 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf"] Dec 05 10:30:06 crc kubenswrapper[4815]: I1205 10:30:06.478262 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415465-dcxmf"] Dec 05 10:30:07 crc kubenswrapper[4815]: I1205 10:30:07.431534 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d134fdd6-8898-49a1-8b2e-78d65bafd6d7" path="/var/lib/kubelet/pods/d134fdd6-8898-49a1-8b2e-78d65bafd6d7/volumes" Dec 05 10:30:08 crc kubenswrapper[4815]: I1205 10:30:08.029985 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_28f58061-0d04-4513-96f0-7d221940a154/memcached/0.log" Dec 05 10:30:10 crc kubenswrapper[4815]: I1205 10:30:10.336161 4815 scope.go:117] "RemoveContainer" containerID="d934e352a34ce974070c6e3d67fc9faeca7b4495d40edce7819650ea82d899b9" Dec 05 10:30:18 crc kubenswrapper[4815]: I1205 10:30:18.870773 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9mfw2"] Dec 05 10:30:18 crc kubenswrapper[4815]: E1205 10:30:18.871785 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1660774c-b853-4b8e-8e44-e581e3623a67" containerName="collect-profiles" Dec 05 10:30:18 crc kubenswrapper[4815]: I1205 10:30:18.871803 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="1660774c-b853-4b8e-8e44-e581e3623a67" containerName="collect-profiles" Dec 05 10:30:18 crc kubenswrapper[4815]: I1205 10:30:18.872013 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="1660774c-b853-4b8e-8e44-e581e3623a67" containerName="collect-profiles" Dec 05 10:30:18 crc kubenswrapper[4815]: I1205 10:30:18.878622 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:18 crc kubenswrapper[4815]: I1205 10:30:18.896063 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9mfw2"] Dec 05 10:30:19 crc kubenswrapper[4815]: I1205 10:30:19.005158 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxjn2\" (UniqueName: \"kubernetes.io/projected/127e1dd1-cc68-4b45-8ec4-c3cab5989682-kube-api-access-cxjn2\") pod \"community-operators-9mfw2\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:19 crc kubenswrapper[4815]: I1205 10:30:19.005591 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-catalog-content\") pod \"community-operators-9mfw2\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:19 crc kubenswrapper[4815]: I1205 10:30:19.005686 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-utilities\") pod \"community-operators-9mfw2\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:19 crc kubenswrapper[4815]: I1205 10:30:19.107612 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxjn2\" (UniqueName: \"kubernetes.io/projected/127e1dd1-cc68-4b45-8ec4-c3cab5989682-kube-api-access-cxjn2\") pod \"community-operators-9mfw2\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:19 crc kubenswrapper[4815]: I1205 10:30:19.107704 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-catalog-content\") pod \"community-operators-9mfw2\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:19 crc kubenswrapper[4815]: I1205 10:30:19.107774 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-utilities\") pod \"community-operators-9mfw2\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:19 crc kubenswrapper[4815]: I1205 10:30:19.108401 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-catalog-content\") pod \"community-operators-9mfw2\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:19 crc kubenswrapper[4815]: I1205 10:30:19.108455 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-utilities\") pod \"community-operators-9mfw2\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:19 crc kubenswrapper[4815]: I1205 10:30:19.130062 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxjn2\" (UniqueName: \"kubernetes.io/projected/127e1dd1-cc68-4b45-8ec4-c3cab5989682-kube-api-access-cxjn2\") pod \"community-operators-9mfw2\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:19 crc kubenswrapper[4815]: I1205 10:30:19.244115 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:19 crc kubenswrapper[4815]: I1205 10:30:19.775881 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9mfw2"] Dec 05 10:30:20 crc kubenswrapper[4815]: I1205 10:30:20.431501 4815 generic.go:334] "Generic (PLEG): container finished" podID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" containerID="0f48473724a92e43b9909aff532fb1aa9f06f735b24328c045906d7224a63a20" exitCode=0 Dec 05 10:30:20 crc kubenswrapper[4815]: I1205 10:30:20.431542 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mfw2" event={"ID":"127e1dd1-cc68-4b45-8ec4-c3cab5989682","Type":"ContainerDied","Data":"0f48473724a92e43b9909aff532fb1aa9f06f735b24328c045906d7224a63a20"} Dec 05 10:30:20 crc kubenswrapper[4815]: I1205 10:30:20.431567 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mfw2" event={"ID":"127e1dd1-cc68-4b45-8ec4-c3cab5989682","Type":"ContainerStarted","Data":"1c3b66048f7cd7c8e4e1531b590ede2e9d89252dac8ae480f4a563670dd3f90e"} Dec 05 10:30:21 crc kubenswrapper[4815]: I1205 10:30:21.441539 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mfw2" event={"ID":"127e1dd1-cc68-4b45-8ec4-c3cab5989682","Type":"ContainerStarted","Data":"c6b6cc5e09786486ae3157a60925241d64e31f4d3155afde19dccf5f245d04e4"} Dec 05 10:30:22 crc kubenswrapper[4815]: I1205 10:30:22.450872 4815 generic.go:334] "Generic (PLEG): container finished" podID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" containerID="c6b6cc5e09786486ae3157a60925241d64e31f4d3155afde19dccf5f245d04e4" exitCode=0 Dec 05 10:30:22 crc kubenswrapper[4815]: I1205 10:30:22.450922 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mfw2" event={"ID":"127e1dd1-cc68-4b45-8ec4-c3cab5989682","Type":"ContainerDied","Data":"c6b6cc5e09786486ae3157a60925241d64e31f4d3155afde19dccf5f245d04e4"} Dec 05 10:30:24 crc kubenswrapper[4815]: I1205 10:30:24.481237 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mfw2" event={"ID":"127e1dd1-cc68-4b45-8ec4-c3cab5989682","Type":"ContainerStarted","Data":"b0b8c1c7e7208af183bc5af8c90b8e8038c712dd6cef8d3fa893c3421efb9ee3"} Dec 05 10:30:24 crc kubenswrapper[4815]: I1205 10:30:24.505338 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9mfw2" podStartSLOduration=3.890338923 podStartE2EDuration="6.505323771s" podCreationTimestamp="2025-12-05 10:30:18 +0000 UTC" firstStartedPulling="2025-12-05 10:30:20.433858836 +0000 UTC m=+5019.312465673" lastFinishedPulling="2025-12-05 10:30:23.048843684 +0000 UTC m=+5021.927450521" observedRunningTime="2025-12-05 10:30:24.502709371 +0000 UTC m=+5023.381316208" watchObservedRunningTime="2025-12-05 10:30:24.505323771 +0000 UTC m=+5023.383930608" Dec 05 10:30:27 crc kubenswrapper[4815]: I1205 10:30:27.180389 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/util/0.log" Dec 05 10:30:27 crc kubenswrapper[4815]: I1205 10:30:27.331466 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/pull/0.log" Dec 05 10:30:27 crc kubenswrapper[4815]: I1205 10:30:27.367803 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/util/0.log" Dec 05 10:30:27 crc kubenswrapper[4815]: I1205 10:30:27.438558 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/pull/0.log" Dec 05 10:30:27 crc kubenswrapper[4815]: I1205 10:30:27.697133 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/pull/0.log" Dec 05 10:30:27 crc kubenswrapper[4815]: I1205 10:30:27.764587 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/extract/0.log" Dec 05 10:30:27 crc kubenswrapper[4815]: I1205 10:30:27.817153 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/util/0.log" Dec 05 10:30:27 crc kubenswrapper[4815]: I1205 10:30:27.944445 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-26tt9_cafb574b-a659-41e6-89d5-c02851274f73/kube-rbac-proxy/0.log" Dec 05 10:30:28 crc kubenswrapper[4815]: I1205 10:30:28.127821 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-j8cxc_f023493d-1716-40ac-816d-26067f0019f3/kube-rbac-proxy/0.log" Dec 05 10:30:28 crc kubenswrapper[4815]: I1205 10:30:28.131866 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-26tt9_cafb574b-a659-41e6-89d5-c02851274f73/manager/0.log" Dec 05 10:30:28 crc kubenswrapper[4815]: I1205 10:30:28.360338 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-g4nbp_9db099f9-ed44-46ed-ab3d-97d09ee0b5b0/kube-rbac-proxy/0.log" Dec 05 10:30:28 crc kubenswrapper[4815]: I1205 10:30:28.597345 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-8blh4_591bafca-5706-418f-b803-d29d1b0865db/kube-rbac-proxy/0.log" Dec 05 10:30:28 crc kubenswrapper[4815]: I1205 10:30:28.923692 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-g4nbp_9db099f9-ed44-46ed-ab3d-97d09ee0b5b0/manager/0.log" Dec 05 10:30:28 crc kubenswrapper[4815]: I1205 10:30:28.975256 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-j8cxc_f023493d-1716-40ac-816d-26067f0019f3/manager/0.log" Dec 05 10:30:28 crc kubenswrapper[4815]: I1205 10:30:28.996602 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-8blh4_591bafca-5706-418f-b803-d29d1b0865db/manager/0.log" Dec 05 10:30:29 crc kubenswrapper[4815]: I1205 10:30:29.025249 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-tmg9c_f23f3b0d-7874-4906-81de-b8d9226082d2/kube-rbac-proxy/0.log" Dec 05 10:30:29 crc kubenswrapper[4815]: I1205 10:30:29.130828 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-tmg9c_f23f3b0d-7874-4906-81de-b8d9226082d2/manager/0.log" Dec 05 10:30:29 crc kubenswrapper[4815]: I1205 10:30:29.205689 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5jh5q_ed463fa2-6734-41b5-a34b-cb83436130fd/kube-rbac-proxy/0.log" Dec 05 10:30:29 crc kubenswrapper[4815]: I1205 10:30:29.233780 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5jh5q_ed463fa2-6734-41b5-a34b-cb83436130fd/manager/0.log" Dec 05 10:30:29 crc kubenswrapper[4815]: I1205 10:30:29.244667 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:29 crc kubenswrapper[4815]: I1205 10:30:29.244716 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:29 crc kubenswrapper[4815]: I1205 10:30:29.923604 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:30 crc kubenswrapper[4815]: I1205 10:30:30.000103 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:30 crc kubenswrapper[4815]: I1205 10:30:30.116421 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-9pflh_4abce90f-180f-4302-aecd-89bf5739757c/kube-rbac-proxy/0.log" Dec 05 10:30:30 crc kubenswrapper[4815]: I1205 10:30:30.168275 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9mfw2"] Dec 05 10:30:30 crc kubenswrapper[4815]: I1205 10:30:30.281377 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-9pflh_4abce90f-180f-4302-aecd-89bf5739757c/manager/0.log" Dec 05 10:30:30 crc kubenswrapper[4815]: I1205 10:30:30.335939 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7mtff_ef1e93b5-9a31-4588-a5e4-78dbe6ef0437/kube-rbac-proxy/0.log" Dec 05 10:30:30 crc kubenswrapper[4815]: I1205 10:30:30.409121 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7mtff_ef1e93b5-9a31-4588-a5e4-78dbe6ef0437/manager/0.log" Dec 05 10:30:30 crc kubenswrapper[4815]: I1205 10:30:30.495006 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-w7vjm_cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb/kube-rbac-proxy/0.log" Dec 05 10:30:30 crc kubenswrapper[4815]: I1205 10:30:30.689803 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-w7vjm_cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb/manager/0.log" Dec 05 10:30:30 crc kubenswrapper[4815]: I1205 10:30:30.755820 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-2k8wt_f80008bd-769f-4960-be8f-62894fdd7718/kube-rbac-proxy/0.log" Dec 05 10:30:30 crc kubenswrapper[4815]: I1205 10:30:30.855103 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-2k8wt_f80008bd-769f-4960-be8f-62894fdd7718/manager/0.log" Dec 05 10:30:30 crc kubenswrapper[4815]: I1205 10:30:30.922663 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-bm8kq_3665adde-25ad-4261-be50-ef4e21726b7b/kube-rbac-proxy/0.log" Dec 05 10:30:31 crc kubenswrapper[4815]: I1205 10:30:31.073483 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-bm8kq_3665adde-25ad-4261-be50-ef4e21726b7b/manager/0.log" Dec 05 10:30:31 crc kubenswrapper[4815]: I1205 10:30:31.549265 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9mfw2" podUID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" containerName="registry-server" containerID="cri-o://b0b8c1c7e7208af183bc5af8c90b8e8038c712dd6cef8d3fa893c3421efb9ee3" gracePeriod=2 Dec 05 10:30:31 crc kubenswrapper[4815]: I1205 10:30:31.626904 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-ccd95_7482e78f-1387-4e91-b5d4-419f5164fea5/kube-rbac-proxy/0.log" Dec 05 10:30:31 crc kubenswrapper[4815]: I1205 10:30:31.672192 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-ccd95_7482e78f-1387-4e91-b5d4-419f5164fea5/manager/0.log" Dec 05 10:30:31 crc kubenswrapper[4815]: I1205 10:30:31.903410 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-25lnj_48f9720a-5d73-4773-85f9-9c571c9322ee/kube-rbac-proxy/0.log" Dec 05 10:30:31 crc kubenswrapper[4815]: I1205 10:30:31.948905 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-l5qqd_b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d/kube-rbac-proxy/0.log" Dec 05 10:30:31 crc kubenswrapper[4815]: I1205 10:30:31.950716 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-l5qqd_b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d/manager/0.log" Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.048929 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-25lnj_48f9720a-5d73-4773-85f9-9c571c9322ee/manager/0.log" Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.192279 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55c85496f5b4xkh_b7558b70-107c-48ea-ac22-e42b1bcdf47d/kube-rbac-proxy/0.log" Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.213021 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55c85496f5b4xkh_b7558b70-107c-48ea-ac22-e42b1bcdf47d/manager/0.log" Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.599075 4815 generic.go:334] "Generic (PLEG): container finished" podID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" containerID="b0b8c1c7e7208af183bc5af8c90b8e8038c712dd6cef8d3fa893c3421efb9ee3" exitCode=0 Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.599127 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mfw2" event={"ID":"127e1dd1-cc68-4b45-8ec4-c3cab5989682","Type":"ContainerDied","Data":"b0b8c1c7e7208af183bc5af8c90b8e8038c712dd6cef8d3fa893c3421efb9ee3"} Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.740958 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.767845 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xttqw_6943ef97-36ce-4f74-b31b-4ba90192d3ac/registry-server/0.log" Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.827637 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-55b6fb9447-xg6vd_7f520180-b5ab-420b-9479-4089e1845ecc/operator/0.log" Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.849551 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-5jlqr_378923e7-18f6-4946-8e86-d1ececca15cc/kube-rbac-proxy/0.log" Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.903936 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxjn2\" (UniqueName: \"kubernetes.io/projected/127e1dd1-cc68-4b45-8ec4-c3cab5989682-kube-api-access-cxjn2\") pod \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.904136 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-catalog-content\") pod \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.904170 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-utilities\") pod \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\" (UID: \"127e1dd1-cc68-4b45-8ec4-c3cab5989682\") " Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.905862 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-utilities" (OuterVolumeSpecName: "utilities") pod "127e1dd1-cc68-4b45-8ec4-c3cab5989682" (UID: "127e1dd1-cc68-4b45-8ec4-c3cab5989682"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.925893 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/127e1dd1-cc68-4b45-8ec4-c3cab5989682-kube-api-access-cxjn2" (OuterVolumeSpecName: "kube-api-access-cxjn2") pod "127e1dd1-cc68-4b45-8ec4-c3cab5989682" (UID: "127e1dd1-cc68-4b45-8ec4-c3cab5989682"). InnerVolumeSpecName "kube-api-access-cxjn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:30:32 crc kubenswrapper[4815]: I1205 10:30:32.967749 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "127e1dd1-cc68-4b45-8ec4-c3cab5989682" (UID: "127e1dd1-cc68-4b45-8ec4-c3cab5989682"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.006672 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.006707 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/127e1dd1-cc68-4b45-8ec4-c3cab5989682-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.006717 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxjn2\" (UniqueName: \"kubernetes.io/projected/127e1dd1-cc68-4b45-8ec4-c3cab5989682-kube-api-access-cxjn2\") on node \"crc\" DevicePath \"\"" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.107113 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-5jlqr_378923e7-18f6-4946-8e86-d1ececca15cc/manager/0.log" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.139101 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-8pt5c_6b41117c-2922-4337-b00c-2bebd57a2c6b/kube-rbac-proxy/0.log" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.248225 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-8pt5c_6b41117c-2922-4337-b00c-2bebd57a2c6b/manager/0.log" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.453831 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mvkws_888d282d-8aa1-4a36-9286-d00af63950ba/operator/0.log" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.471583 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6nl6f_47664970-bfba-453a-ad06-ac1e03417eac/kube-rbac-proxy/0.log" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.613925 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9mfw2" event={"ID":"127e1dd1-cc68-4b45-8ec4-c3cab5989682","Type":"ContainerDied","Data":"1c3b66048f7cd7c8e4e1531b590ede2e9d89252dac8ae480f4a563670dd3f90e"} Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.614282 4815 scope.go:117] "RemoveContainer" containerID="b0b8c1c7e7208af183bc5af8c90b8e8038c712dd6cef8d3fa893c3421efb9ee3" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.614445 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9mfw2" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.655095 4815 scope.go:117] "RemoveContainer" containerID="c6b6cc5e09786486ae3157a60925241d64e31f4d3155afde19dccf5f245d04e4" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.655608 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9mfw2"] Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.671132 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9mfw2"] Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.697526 4815 scope.go:117] "RemoveContainer" containerID="0f48473724a92e43b9909aff532fb1aa9f06f735b24328c045906d7224a63a20" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.817602 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6nl6f_47664970-bfba-453a-ad06-ac1e03417eac/manager/0.log" Dec 05 10:30:33 crc kubenswrapper[4815]: I1205 10:30:33.975311 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bnb4c_926f3f91-cee1-42ae-888f-73ec7f20dd61/kube-rbac-proxy/0.log" Dec 05 10:30:34 crc kubenswrapper[4815]: I1205 10:30:34.021970 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-54bdf956c4-2lknm_8f0cbf2b-67d6-4058-b173-d24117dfcdd7/manager/0.log" Dec 05 10:30:34 crc kubenswrapper[4815]: I1205 10:30:34.076334 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bnb4c_926f3f91-cee1-42ae-888f-73ec7f20dd61/manager/0.log" Dec 05 10:30:34 crc kubenswrapper[4815]: I1205 10:30:34.200354 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-gbkbf_d8b41c8b-030e-4082-98a3-a0e9fc05a208/kube-rbac-proxy/0.log" Dec 05 10:30:34 crc kubenswrapper[4815]: I1205 10:30:34.218252 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-gbkbf_d8b41c8b-030e-4082-98a3-a0e9fc05a208/manager/0.log" Dec 05 10:30:34 crc kubenswrapper[4815]: I1205 10:30:34.372613 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-j4r7f_19f20867-5315-4be3-836b-57bb0b501b36/kube-rbac-proxy/0.log" Dec 05 10:30:34 crc kubenswrapper[4815]: I1205 10:30:34.522709 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-j4r7f_19f20867-5315-4be3-836b-57bb0b501b36/manager/0.log" Dec 05 10:30:35 crc kubenswrapper[4815]: I1205 10:30:35.431684 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" path="/var/lib/kubelet/pods/127e1dd1-cc68-4b45-8ec4-c3cab5989682/volumes" Dec 05 10:30:50 crc kubenswrapper[4815]: I1205 10:30:50.207229 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:30:50 crc kubenswrapper[4815]: I1205 10:30:50.207920 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:30:56 crc kubenswrapper[4815]: I1205 10:30:56.827480 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-h89pl_fb48516f-711a-4131-a555-de5ee6e96ae1/control-plane-machine-set-operator/0.log" Dec 05 10:30:57 crc kubenswrapper[4815]: I1205 10:30:57.041321 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tw877_afd36151-d969-4642-b6c4-c6b3b5f65476/kube-rbac-proxy/0.log" Dec 05 10:30:57 crc kubenswrapper[4815]: I1205 10:30:57.063830 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tw877_afd36151-d969-4642-b6c4-c6b3b5f65476/machine-api-operator/0.log" Dec 05 10:31:12 crc kubenswrapper[4815]: I1205 10:31:12.642623 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vcstd_82c69bfe-d893-41ec-bc5d-d872d51c8c15/cert-manager-controller/0.log" Dec 05 10:31:12 crc kubenswrapper[4815]: I1205 10:31:12.930237 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-zc6mp_8f3da4f1-671e-4911-a481-45824ce433d5/cert-manager-cainjector/0.log" Dec 05 10:31:13 crc kubenswrapper[4815]: I1205 10:31:13.035753 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-sz66k_56e3935a-c7dd-4bb2-9dff-deaf3dd7e25b/cert-manager-webhook/0.log" Dec 05 10:31:20 crc kubenswrapper[4815]: I1205 10:31:20.192832 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:31:20 crc kubenswrapper[4815]: I1205 10:31:20.197051 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:31:29 crc kubenswrapper[4815]: I1205 10:31:29.492424 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-5g45b_028ef3e8-a0e1-443e-840d-0f7244a0e953/nmstate-console-plugin/0.log" Dec 05 10:31:29 crc kubenswrapper[4815]: I1205 10:31:29.655730 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-d8hkx_da8c41e3-2e9f-4917-a9e7-efa4debc22c7/nmstate-handler/0.log" Dec 05 10:31:29 crc kubenswrapper[4815]: I1205 10:31:29.686381 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9frbq_d47f3938-b642-4bd2-b807-b2c962918651/kube-rbac-proxy/0.log" Dec 05 10:31:29 crc kubenswrapper[4815]: I1205 10:31:29.733625 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9frbq_d47f3938-b642-4bd2-b807-b2c962918651/nmstate-metrics/0.log" Dec 05 10:31:29 crc kubenswrapper[4815]: I1205 10:31:29.912534 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-82r92_7e7d72c9-18b4-4716-aabb-2c55c7070020/nmstate-operator/0.log" Dec 05 10:31:30 crc kubenswrapper[4815]: I1205 10:31:30.027296 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-qtprq_8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5/nmstate-webhook/0.log" Dec 05 10:31:48 crc kubenswrapper[4815]: I1205 10:31:48.202385 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-j8cbz_55c34ec0-bd62-4eb5-a838-48e0622ebea9/kube-rbac-proxy/0.log" Dec 05 10:31:48 crc kubenswrapper[4815]: I1205 10:31:48.450876 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-j8cbz_55c34ec0-bd62-4eb5-a838-48e0622ebea9/controller/0.log" Dec 05 10:31:48 crc kubenswrapper[4815]: I1205 10:31:48.522771 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-frr-files/0.log" Dec 05 10:31:49 crc kubenswrapper[4815]: I1205 10:31:49.053776 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-frr-files/0.log" Dec 05 10:31:49 crc kubenswrapper[4815]: I1205 10:31:49.295022 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-reloader/0.log" Dec 05 10:31:49 crc kubenswrapper[4815]: I1205 10:31:49.359773 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-metrics/0.log" Dec 05 10:31:49 crc kubenswrapper[4815]: I1205 10:31:49.359981 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-reloader/0.log" Dec 05 10:31:49 crc kubenswrapper[4815]: I1205 10:31:49.711120 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-reloader/0.log" Dec 05 10:31:49 crc kubenswrapper[4815]: I1205 10:31:49.712314 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-metrics/0.log" Dec 05 10:31:49 crc kubenswrapper[4815]: I1205 10:31:49.720719 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-frr-files/0.log" Dec 05 10:31:49 crc kubenswrapper[4815]: I1205 10:31:49.803946 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-metrics/0.log" Dec 05 10:31:49 crc kubenswrapper[4815]: I1205 10:31:49.969578 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-metrics/0.log" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.012505 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-frr-files/0.log" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.042080 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/controller/0.log" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.042165 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-reloader/0.log" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.193008 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.193267 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.193384 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.194252 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.194387 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" gracePeriod=600 Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.234291 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/frr-metrics/0.log" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.299691 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/kube-rbac-proxy-frr/0.log" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.314574 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/kube-rbac-proxy/0.log" Dec 05 10:31:50 crc kubenswrapper[4815]: E1205 10:31:50.340553 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.504629 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" exitCode=0 Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.504686 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f"} Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.504726 4815 scope.go:117] "RemoveContainer" containerID="d34cc10d62df8ab9e8c19668427c9711c21dc528084d7e09c84a9817ac5c73f0" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.505588 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:31:50 crc kubenswrapper[4815]: E1205 10:31:50.505931 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.603630 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/reloader/0.log" Dec 05 10:31:50 crc kubenswrapper[4815]: I1205 10:31:50.739345 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-b6cnn_46274b62-2000-4c62-8688-af6b5fcb87dc/frr-k8s-webhook-server/0.log" Dec 05 10:31:51 crc kubenswrapper[4815]: I1205 10:31:51.021563 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7f5859557b-zxp74_37e3bb94-27f2-4bfe-9c26-4bbb3025330c/manager/0.log" Dec 05 10:31:51 crc kubenswrapper[4815]: I1205 10:31:51.346410 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-69858568cc-scmsp_1aabbb7d-8736-4ff3-a41e-7599a4e74194/webhook-server/0.log" Dec 05 10:31:51 crc kubenswrapper[4815]: I1205 10:31:51.467071 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xszxj_a72253a0-2394-424d-8689-1c989ce6b2ed/kube-rbac-proxy/0.log" Dec 05 10:31:51 crc kubenswrapper[4815]: I1205 10:31:51.786592 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/frr/0.log" Dec 05 10:31:51 crc kubenswrapper[4815]: I1205 10:31:51.950929 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xszxj_a72253a0-2394-424d-8689-1c989ce6b2ed/speaker/0.log" Dec 05 10:32:02 crc kubenswrapper[4815]: I1205 10:32:02.418754 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:32:02 crc kubenswrapper[4815]: E1205 10:32:02.419678 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:32:09 crc kubenswrapper[4815]: I1205 10:32:09.055160 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/util/0.log" Dec 05 10:32:09 crc kubenswrapper[4815]: I1205 10:32:09.256006 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/util/0.log" Dec 05 10:32:09 crc kubenswrapper[4815]: I1205 10:32:09.302866 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/pull/0.log" Dec 05 10:32:09 crc kubenswrapper[4815]: I1205 10:32:09.318003 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/pull/0.log" Dec 05 10:32:09 crc kubenswrapper[4815]: I1205 10:32:09.542949 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/extract/0.log" Dec 05 10:32:09 crc kubenswrapper[4815]: I1205 10:32:09.567739 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/util/0.log" Dec 05 10:32:09 crc kubenswrapper[4815]: I1205 10:32:09.587738 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/pull/0.log" Dec 05 10:32:09 crc kubenswrapper[4815]: I1205 10:32:09.821723 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/util/0.log" Dec 05 10:32:09 crc kubenswrapper[4815]: I1205 10:32:09.992162 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/util/0.log" Dec 05 10:32:10 crc kubenswrapper[4815]: I1205 10:32:10.034581 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/pull/0.log" Dec 05 10:32:10 crc kubenswrapper[4815]: I1205 10:32:10.072651 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/pull/0.log" Dec 05 10:32:10 crc kubenswrapper[4815]: I1205 10:32:10.261644 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/extract/0.log" Dec 05 10:32:10 crc kubenswrapper[4815]: I1205 10:32:10.289730 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/util/0.log" Dec 05 10:32:10 crc kubenswrapper[4815]: I1205 10:32:10.295038 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/pull/0.log" Dec 05 10:32:10 crc kubenswrapper[4815]: I1205 10:32:10.515125 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-utilities/0.log" Dec 05 10:32:10 crc kubenswrapper[4815]: I1205 10:32:10.776286 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-content/0.log" Dec 05 10:32:10 crc kubenswrapper[4815]: I1205 10:32:10.785513 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-utilities/0.log" Dec 05 10:32:10 crc kubenswrapper[4815]: I1205 10:32:10.799501 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-content/0.log" Dec 05 10:32:11 crc kubenswrapper[4815]: I1205 10:32:11.044361 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-content/0.log" Dec 05 10:32:11 crc kubenswrapper[4815]: I1205 10:32:11.088262 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-utilities/0.log" Dec 05 10:32:11 crc kubenswrapper[4815]: I1205 10:32:11.331632 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-utilities/0.log" Dec 05 10:32:11 crc kubenswrapper[4815]: I1205 10:32:11.767135 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/registry-server/0.log" Dec 05 10:32:11 crc kubenswrapper[4815]: I1205 10:32:11.896870 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-utilities/0.log" Dec 05 10:32:11 crc kubenswrapper[4815]: I1205 10:32:11.979200 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-content/0.log" Dec 05 10:32:12 crc kubenswrapper[4815]: I1205 10:32:12.007829 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-content/0.log" Dec 05 10:32:12 crc kubenswrapper[4815]: I1205 10:32:12.242613 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-content/0.log" Dec 05 10:32:12 crc kubenswrapper[4815]: I1205 10:32:12.267520 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-utilities/0.log" Dec 05 10:32:12 crc kubenswrapper[4815]: I1205 10:32:12.644868 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-utilities/0.log" Dec 05 10:32:12 crc kubenswrapper[4815]: I1205 10:32:12.784074 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6k7mn_a4c4b634-bebd-41c4-9462-d33c8a9ff1cd/marketplace-operator/0.log" Dec 05 10:32:13 crc kubenswrapper[4815]: I1205 10:32:13.222412 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/registry-server/0.log" Dec 05 10:32:13 crc kubenswrapper[4815]: I1205 10:32:13.384544 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-content/0.log" Dec 05 10:32:13 crc kubenswrapper[4815]: I1205 10:32:13.436988 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-utilities/0.log" Dec 05 10:32:13 crc kubenswrapper[4815]: I1205 10:32:13.497831 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-content/0.log" Dec 05 10:32:13 crc kubenswrapper[4815]: I1205 10:32:13.614124 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-content/0.log" Dec 05 10:32:13 crc kubenswrapper[4815]: I1205 10:32:13.687134 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-utilities/0.log" Dec 05 10:32:13 crc kubenswrapper[4815]: I1205 10:32:13.807252 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/registry-server/0.log" Dec 05 10:32:13 crc kubenswrapper[4815]: I1205 10:32:13.816346 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-utilities/0.log" Dec 05 10:32:14 crc kubenswrapper[4815]: I1205 10:32:14.086613 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-utilities/0.log" Dec 05 10:32:14 crc kubenswrapper[4815]: I1205 10:32:14.116449 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-content/0.log" Dec 05 10:32:14 crc kubenswrapper[4815]: I1205 10:32:14.127097 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-content/0.log" Dec 05 10:32:14 crc kubenswrapper[4815]: I1205 10:32:14.329232 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-content/0.log" Dec 05 10:32:14 crc kubenswrapper[4815]: I1205 10:32:14.386466 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-utilities/0.log" Dec 05 10:32:14 crc kubenswrapper[4815]: I1205 10:32:14.418931 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:32:14 crc kubenswrapper[4815]: E1205 10:32:14.419201 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:32:15 crc kubenswrapper[4815]: I1205 10:32:15.169431 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/registry-server/0.log" Dec 05 10:32:29 crc kubenswrapper[4815]: I1205 10:32:29.419722 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:32:29 crc kubenswrapper[4815]: E1205 10:32:29.420666 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:32:37 crc kubenswrapper[4815]: E1205 10:32:37.092117 4815 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.129.56.67:57066->38.129.56.67:45709: write tcp 38.129.56.67:57066->38.129.56.67:45709: write: broken pipe Dec 05 10:32:44 crc kubenswrapper[4815]: I1205 10:32:44.419768 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:32:44 crc kubenswrapper[4815]: E1205 10:32:44.420394 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:32:59 crc kubenswrapper[4815]: I1205 10:32:59.419134 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:32:59 crc kubenswrapper[4815]: E1205 10:32:59.420303 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:33:13 crc kubenswrapper[4815]: I1205 10:33:13.418944 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:33:13 crc kubenswrapper[4815]: E1205 10:33:13.419736 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:33:28 crc kubenswrapper[4815]: I1205 10:33:28.419042 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:33:28 crc kubenswrapper[4815]: E1205 10:33:28.419929 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:33:39 crc kubenswrapper[4815]: I1205 10:33:39.419860 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:33:39 crc kubenswrapper[4815]: E1205 10:33:39.420708 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:33:52 crc kubenswrapper[4815]: I1205 10:33:52.421564 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:33:52 crc kubenswrapper[4815]: E1205 10:33:52.422664 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:33:57 crc kubenswrapper[4815]: I1205 10:33:57.453687 4815 patch_prober.go:28] interesting pod/route-controller-manager-769fd76f7f-bm6zj container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 05 10:33:57 crc kubenswrapper[4815]: I1205 10:33:57.454264 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-769fd76f7f-bm6zj" podUID="8ab1ff45-e50b-4644-bd45-2436a76dae21" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.69:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 05 10:34:04 crc kubenswrapper[4815]: I1205 10:34:04.828193 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="078b665f-5e54-4d83-966b-3684d06ea320" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Dec 05 10:34:04 crc kubenswrapper[4815]: I1205 10:34:04.871617 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-jbdfv" podUID="563ef3a0-d97f-4b5b-afc4-378bdbbee19b" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 05 10:34:06 crc kubenswrapper[4815]: I1205 10:34:06.419941 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:34:06 crc kubenswrapper[4815]: E1205 10:34:06.421478 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:34:10 crc kubenswrapper[4815]: I1205 10:34:10.487327 4815 scope.go:117] "RemoveContainer" containerID="6f6ad4fa3ed36c0d72c02204da595c93e9b78921a0bd5648b1361054f10ffb59" Dec 05 10:34:10 crc kubenswrapper[4815]: I1205 10:34:10.540140 4815 scope.go:117] "RemoveContainer" containerID="4ff58c55e0ac50c928ea60b59ec9724ba9f03cf007331dc37a3ca34b3973f952" Dec 05 10:34:10 crc kubenswrapper[4815]: I1205 10:34:10.563818 4815 scope.go:117] "RemoveContainer" containerID="e87bbcc2cd6b6156e25ad1cf8eaf9cbe0090651fcfb25c600710d1b6760ccbd3" Dec 05 10:34:10 crc kubenswrapper[4815]: I1205 10:34:10.611905 4815 scope.go:117] "RemoveContainer" containerID="b9c39accfa11678af00a232ec0a62ebd2934f230b428b87b2aff0bc4c9edfab8" Dec 05 10:34:17 crc kubenswrapper[4815]: I1205 10:34:17.420611 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:34:17 crc kubenswrapper[4815]: E1205 10:34:17.421443 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:34:29 crc kubenswrapper[4815]: I1205 10:34:29.423856 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:34:29 crc kubenswrapper[4815]: E1205 10:34:29.424885 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:34:42 crc kubenswrapper[4815]: I1205 10:34:42.419170 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:34:42 crc kubenswrapper[4815]: E1205 10:34:42.420096 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:34:53 crc kubenswrapper[4815]: I1205 10:34:53.419085 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:34:53 crc kubenswrapper[4815]: E1205 10:34:53.420051 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:35:05 crc kubenswrapper[4815]: I1205 10:35:05.414950 4815 generic.go:334] "Generic (PLEG): container finished" podID="edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" containerID="595cff59c7f177648979507968c0b00f42dc7874b2607fe12f56d5f62dfea3da" exitCode=0 Dec 05 10:35:05 crc kubenswrapper[4815]: I1205 10:35:05.415035 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-w2mx6/must-gather-gsjwp" event={"ID":"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2","Type":"ContainerDied","Data":"595cff59c7f177648979507968c0b00f42dc7874b2607fe12f56d5f62dfea3da"} Dec 05 10:35:05 crc kubenswrapper[4815]: I1205 10:35:05.417456 4815 scope.go:117] "RemoveContainer" containerID="595cff59c7f177648979507968c0b00f42dc7874b2607fe12f56d5f62dfea3da" Dec 05 10:35:06 crc kubenswrapper[4815]: I1205 10:35:06.354018 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w2mx6_must-gather-gsjwp_edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2/gather/0.log" Dec 05 10:35:07 crc kubenswrapper[4815]: I1205 10:35:07.423773 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:35:07 crc kubenswrapper[4815]: E1205 10:35:07.424505 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:35:16 crc kubenswrapper[4815]: I1205 10:35:16.802168 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-w2mx6/must-gather-gsjwp"] Dec 05 10:35:16 crc kubenswrapper[4815]: I1205 10:35:16.802911 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-w2mx6/must-gather-gsjwp" podUID="edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" containerName="copy" containerID="cri-o://e6ef2f836c755fb6720d5d69ca50e83d05646c9e0badb80bc694ea53f50bc7f6" gracePeriod=2 Dec 05 10:35:16 crc kubenswrapper[4815]: I1205 10:35:16.816015 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-w2mx6/must-gather-gsjwp"] Dec 05 10:35:17 crc kubenswrapper[4815]: I1205 10:35:17.564879 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w2mx6_must-gather-gsjwp_edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2/copy/0.log" Dec 05 10:35:17 crc kubenswrapper[4815]: I1205 10:35:17.565347 4815 generic.go:334] "Generic (PLEG): container finished" podID="edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" containerID="e6ef2f836c755fb6720d5d69ca50e83d05646c9e0badb80bc694ea53f50bc7f6" exitCode=143 Dec 05 10:35:18 crc kubenswrapper[4815]: I1205 10:35:18.821074 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w2mx6_must-gather-gsjwp_edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2/copy/0.log" Dec 05 10:35:18 crc kubenswrapper[4815]: I1205 10:35:18.825756 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/must-gather-gsjwp" Dec 05 10:35:18 crc kubenswrapper[4815]: I1205 10:35:18.870193 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-must-gather-output\") pod \"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2\" (UID: \"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2\") " Dec 05 10:35:18 crc kubenswrapper[4815]: I1205 10:35:18.870365 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlpxc\" (UniqueName: \"kubernetes.io/projected/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-kube-api-access-qlpxc\") pod \"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2\" (UID: \"edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2\") " Dec 05 10:35:18 crc kubenswrapper[4815]: I1205 10:35:18.881086 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-kube-api-access-qlpxc" (OuterVolumeSpecName: "kube-api-access-qlpxc") pod "edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" (UID: "edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2"). InnerVolumeSpecName "kube-api-access-qlpxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:35:18 crc kubenswrapper[4815]: I1205 10:35:18.971914 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlpxc\" (UniqueName: \"kubernetes.io/projected/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-kube-api-access-qlpxc\") on node \"crc\" DevicePath \"\"" Dec 05 10:35:19 crc kubenswrapper[4815]: I1205 10:35:19.041271 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" (UID: "edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:35:19 crc kubenswrapper[4815]: I1205 10:35:19.074193 4815 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 05 10:35:19 crc kubenswrapper[4815]: I1205 10:35:19.439557 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" path="/var/lib/kubelet/pods/edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2/volumes" Dec 05 10:35:19 crc kubenswrapper[4815]: I1205 10:35:19.604569 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-w2mx6_must-gather-gsjwp_edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2/copy/0.log" Dec 05 10:35:19 crc kubenswrapper[4815]: I1205 10:35:19.605177 4815 scope.go:117] "RemoveContainer" containerID="e6ef2f836c755fb6720d5d69ca50e83d05646c9e0badb80bc694ea53f50bc7f6" Dec 05 10:35:19 crc kubenswrapper[4815]: I1205 10:35:19.605355 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-w2mx6/must-gather-gsjwp" Dec 05 10:35:19 crc kubenswrapper[4815]: I1205 10:35:19.660832 4815 scope.go:117] "RemoveContainer" containerID="595cff59c7f177648979507968c0b00f42dc7874b2607fe12f56d5f62dfea3da" Dec 05 10:35:21 crc kubenswrapper[4815]: I1205 10:35:21.427728 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:35:21 crc kubenswrapper[4815]: E1205 10:35:21.428326 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:35:34 crc kubenswrapper[4815]: I1205 10:35:34.418679 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:35:34 crc kubenswrapper[4815]: E1205 10:35:34.419568 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:35:47 crc kubenswrapper[4815]: I1205 10:35:47.419267 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:35:47 crc kubenswrapper[4815]: E1205 10:35:47.419969 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:35:59 crc kubenswrapper[4815]: I1205 10:35:59.419246 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:35:59 crc kubenswrapper[4815]: E1205 10:35:59.420428 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:36:12 crc kubenswrapper[4815]: I1205 10:36:12.418787 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:36:12 crc kubenswrapper[4815]: E1205 10:36:12.419408 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:36:26 crc kubenswrapper[4815]: I1205 10:36:26.419526 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:36:26 crc kubenswrapper[4815]: E1205 10:36:26.420317 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:36:41 crc kubenswrapper[4815]: I1205 10:36:41.488802 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:36:41 crc kubenswrapper[4815]: E1205 10:36:41.490788 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:36:52 crc kubenswrapper[4815]: I1205 10:36:52.419254 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:36:53 crc kubenswrapper[4815]: I1205 10:36:53.215116 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"36532cd36fe5a7fe6381b37918506e1e1d6db7ef1721fc2254713e8eb9050c93"} Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.445420 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lq64j/must-gather-ttd7k"] Dec 05 10:38:25 crc kubenswrapper[4815]: E1205 10:38:25.446234 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" containerName="extract-content" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.446258 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" containerName="extract-content" Dec 05 10:38:25 crc kubenswrapper[4815]: E1205 10:38:25.446278 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" containerName="extract-utilities" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.446285 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" containerName="extract-utilities" Dec 05 10:38:25 crc kubenswrapper[4815]: E1205 10:38:25.446327 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" containerName="copy" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.446333 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" containerName="copy" Dec 05 10:38:25 crc kubenswrapper[4815]: E1205 10:38:25.446357 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" containerName="gather" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.446363 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" containerName="gather" Dec 05 10:38:25 crc kubenswrapper[4815]: E1205 10:38:25.446376 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" containerName="registry-server" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.446381 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" containerName="registry-server" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.447823 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" containerName="gather" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.447848 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="127e1dd1-cc68-4b45-8ec4-c3cab5989682" containerName="registry-server" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.447864 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="edc9832b-b2a9-464d-a5dc-8e9ec1ec65e2" containerName="copy" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.448919 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/must-gather-ttd7k" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.456816 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lq64j"/"openshift-service-ca.crt" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.456816 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-lq64j"/"default-dockercfg-mrsfg" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.457458 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-lq64j"/"kube-root-ca.crt" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.459952 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-must-gather-output\") pod \"must-gather-ttd7k\" (UID: \"11a003bf-f178-4f1f-a45a-f13e24f6ce8c\") " pod="openshift-must-gather-lq64j/must-gather-ttd7k" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.460043 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cdc7\" (UniqueName: \"kubernetes.io/projected/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-kube-api-access-2cdc7\") pod \"must-gather-ttd7k\" (UID: \"11a003bf-f178-4f1f-a45a-f13e24f6ce8c\") " pod="openshift-must-gather-lq64j/must-gather-ttd7k" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.484741 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lq64j/must-gather-ttd7k"] Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.561721 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-must-gather-output\") pod \"must-gather-ttd7k\" (UID: \"11a003bf-f178-4f1f-a45a-f13e24f6ce8c\") " pod="openshift-must-gather-lq64j/must-gather-ttd7k" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.561838 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cdc7\" (UniqueName: \"kubernetes.io/projected/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-kube-api-access-2cdc7\") pod \"must-gather-ttd7k\" (UID: \"11a003bf-f178-4f1f-a45a-f13e24f6ce8c\") " pod="openshift-must-gather-lq64j/must-gather-ttd7k" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.562220 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-must-gather-output\") pod \"must-gather-ttd7k\" (UID: \"11a003bf-f178-4f1f-a45a-f13e24f6ce8c\") " pod="openshift-must-gather-lq64j/must-gather-ttd7k" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.589256 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cdc7\" (UniqueName: \"kubernetes.io/projected/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-kube-api-access-2cdc7\") pod \"must-gather-ttd7k\" (UID: \"11a003bf-f178-4f1f-a45a-f13e24f6ce8c\") " pod="openshift-must-gather-lq64j/must-gather-ttd7k" Dec 05 10:38:25 crc kubenswrapper[4815]: I1205 10:38:25.771398 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/must-gather-ttd7k" Dec 05 10:38:26 crc kubenswrapper[4815]: I1205 10:38:26.368757 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-lq64j/must-gather-ttd7k"] Dec 05 10:38:27 crc kubenswrapper[4815]: I1205 10:38:27.162365 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lq64j/must-gather-ttd7k" event={"ID":"11a003bf-f178-4f1f-a45a-f13e24f6ce8c","Type":"ContainerStarted","Data":"c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae"} Dec 05 10:38:27 crc kubenswrapper[4815]: I1205 10:38:27.163355 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lq64j/must-gather-ttd7k" event={"ID":"11a003bf-f178-4f1f-a45a-f13e24f6ce8c","Type":"ContainerStarted","Data":"fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be"} Dec 05 10:38:27 crc kubenswrapper[4815]: I1205 10:38:27.163423 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lq64j/must-gather-ttd7k" event={"ID":"11a003bf-f178-4f1f-a45a-f13e24f6ce8c","Type":"ContainerStarted","Data":"78fb05613618f6f126629335e5662fe89d35f49761b8b4f8dc8b932b6b90a18d"} Dec 05 10:38:27 crc kubenswrapper[4815]: I1205 10:38:27.185800 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lq64j/must-gather-ttd7k" podStartSLOduration=2.185749352 podStartE2EDuration="2.185749352s" podCreationTimestamp="2025-12-05 10:38:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:27.180543938 +0000 UTC m=+5506.059150805" watchObservedRunningTime="2025-12-05 10:38:27.185749352 +0000 UTC m=+5506.064356189" Dec 05 10:38:31 crc kubenswrapper[4815]: I1205 10:38:31.455406 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lq64j/crc-debug-9fqqt"] Dec 05 10:38:31 crc kubenswrapper[4815]: I1205 10:38:31.457763 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-9fqqt" Dec 05 10:38:31 crc kubenswrapper[4815]: I1205 10:38:31.482550 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-host\") pod \"crc-debug-9fqqt\" (UID: \"96c22dce-eb0e-42fd-b689-d50a5e68c8fa\") " pod="openshift-must-gather-lq64j/crc-debug-9fqqt" Dec 05 10:38:31 crc kubenswrapper[4815]: I1205 10:38:31.482791 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnkhd\" (UniqueName: \"kubernetes.io/projected/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-kube-api-access-vnkhd\") pod \"crc-debug-9fqqt\" (UID: \"96c22dce-eb0e-42fd-b689-d50a5e68c8fa\") " pod="openshift-must-gather-lq64j/crc-debug-9fqqt" Dec 05 10:38:31 crc kubenswrapper[4815]: I1205 10:38:31.585101 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnkhd\" (UniqueName: \"kubernetes.io/projected/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-kube-api-access-vnkhd\") pod \"crc-debug-9fqqt\" (UID: \"96c22dce-eb0e-42fd-b689-d50a5e68c8fa\") " pod="openshift-must-gather-lq64j/crc-debug-9fqqt" Dec 05 10:38:31 crc kubenswrapper[4815]: I1205 10:38:31.585336 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-host\") pod \"crc-debug-9fqqt\" (UID: \"96c22dce-eb0e-42fd-b689-d50a5e68c8fa\") " pod="openshift-must-gather-lq64j/crc-debug-9fqqt" Dec 05 10:38:31 crc kubenswrapper[4815]: I1205 10:38:31.585511 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-host\") pod \"crc-debug-9fqqt\" (UID: \"96c22dce-eb0e-42fd-b689-d50a5e68c8fa\") " pod="openshift-must-gather-lq64j/crc-debug-9fqqt" Dec 05 10:38:31 crc kubenswrapper[4815]: I1205 10:38:31.606245 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnkhd\" (UniqueName: \"kubernetes.io/projected/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-kube-api-access-vnkhd\") pod \"crc-debug-9fqqt\" (UID: \"96c22dce-eb0e-42fd-b689-d50a5e68c8fa\") " pod="openshift-must-gather-lq64j/crc-debug-9fqqt" Dec 05 10:38:31 crc kubenswrapper[4815]: I1205 10:38:31.779348 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-9fqqt" Dec 05 10:38:32 crc kubenswrapper[4815]: I1205 10:38:32.234256 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lq64j/crc-debug-9fqqt" event={"ID":"96c22dce-eb0e-42fd-b689-d50a5e68c8fa","Type":"ContainerStarted","Data":"17ba3448a891bb2c597210fa2bcd29cedf664200ac5c0bad14979df7325f8f23"} Dec 05 10:38:32 crc kubenswrapper[4815]: I1205 10:38:32.234501 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lq64j/crc-debug-9fqqt" event={"ID":"96c22dce-eb0e-42fd-b689-d50a5e68c8fa","Type":"ContainerStarted","Data":"c387ef227c6058d74662d354d8a0a42aa923df8da547945fc7b645332edd597d"} Dec 05 10:38:32 crc kubenswrapper[4815]: I1205 10:38:32.260346 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-lq64j/crc-debug-9fqqt" podStartSLOduration=1.260325598 podStartE2EDuration="1.260325598s" podCreationTimestamp="2025-12-05 10:38:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-05 10:38:32.251963128 +0000 UTC m=+5511.130569965" watchObservedRunningTime="2025-12-05 10:38:32.260325598 +0000 UTC m=+5511.138932425" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.112423 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rf9tj"] Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.115468 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.131761 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rf9tj"] Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.291320 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-utilities\") pod \"redhat-marketplace-rf9tj\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.291373 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-catalog-content\") pod \"redhat-marketplace-rf9tj\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.291467 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2c29\" (UniqueName: \"kubernetes.io/projected/824220d8-ccda-4316-a037-b7a4ab90bf28-kube-api-access-g2c29\") pod \"redhat-marketplace-rf9tj\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.393221 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2c29\" (UniqueName: \"kubernetes.io/projected/824220d8-ccda-4316-a037-b7a4ab90bf28-kube-api-access-g2c29\") pod \"redhat-marketplace-rf9tj\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.393419 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-utilities\") pod \"redhat-marketplace-rf9tj\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.393451 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-catalog-content\") pod \"redhat-marketplace-rf9tj\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.393984 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-catalog-content\") pod \"redhat-marketplace-rf9tj\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.393993 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-utilities\") pod \"redhat-marketplace-rf9tj\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.435555 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2c29\" (UniqueName: \"kubernetes.io/projected/824220d8-ccda-4316-a037-b7a4ab90bf28-kube-api-access-g2c29\") pod \"redhat-marketplace-rf9tj\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.439606 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:45 crc kubenswrapper[4815]: I1205 10:38:45.978359 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rf9tj"] Dec 05 10:38:46 crc kubenswrapper[4815]: I1205 10:38:46.360280 4815 generic.go:334] "Generic (PLEG): container finished" podID="824220d8-ccda-4316-a037-b7a4ab90bf28" containerID="92520bc899d153147efa8f8d3f7246e0af4e657888f9875845d501cd541b5ab9" exitCode=0 Dec 05 10:38:46 crc kubenswrapper[4815]: I1205 10:38:46.360333 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rf9tj" event={"ID":"824220d8-ccda-4316-a037-b7a4ab90bf28","Type":"ContainerDied","Data":"92520bc899d153147efa8f8d3f7246e0af4e657888f9875845d501cd541b5ab9"} Dec 05 10:38:46 crc kubenswrapper[4815]: I1205 10:38:46.360365 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rf9tj" event={"ID":"824220d8-ccda-4316-a037-b7a4ab90bf28","Type":"ContainerStarted","Data":"a730ddbce109337283181cab8b327cde8a96f6056dc81c4b48323f94bc6c9863"} Dec 05 10:38:46 crc kubenswrapper[4815]: I1205 10:38:46.363292 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:38:47 crc kubenswrapper[4815]: I1205 10:38:47.372096 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rf9tj" event={"ID":"824220d8-ccda-4316-a037-b7a4ab90bf28","Type":"ContainerStarted","Data":"02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83"} Dec 05 10:38:48 crc kubenswrapper[4815]: I1205 10:38:48.381965 4815 generic.go:334] "Generic (PLEG): container finished" podID="824220d8-ccda-4316-a037-b7a4ab90bf28" containerID="02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83" exitCode=0 Dec 05 10:38:48 crc kubenswrapper[4815]: I1205 10:38:48.382117 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rf9tj" event={"ID":"824220d8-ccda-4316-a037-b7a4ab90bf28","Type":"ContainerDied","Data":"02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83"} Dec 05 10:38:49 crc kubenswrapper[4815]: I1205 10:38:49.395867 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rf9tj" event={"ID":"824220d8-ccda-4316-a037-b7a4ab90bf28","Type":"ContainerStarted","Data":"2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489"} Dec 05 10:38:49 crc kubenswrapper[4815]: I1205 10:38:49.430197 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rf9tj" podStartSLOduration=1.9900228389999999 podStartE2EDuration="4.430177394s" podCreationTimestamp="2025-12-05 10:38:45 +0000 UTC" firstStartedPulling="2025-12-05 10:38:46.362109781 +0000 UTC m=+5525.240716618" lastFinishedPulling="2025-12-05 10:38:48.802264336 +0000 UTC m=+5527.680871173" observedRunningTime="2025-12-05 10:38:49.419569341 +0000 UTC m=+5528.298176178" watchObservedRunningTime="2025-12-05 10:38:49.430177394 +0000 UTC m=+5528.308784231" Dec 05 10:38:55 crc kubenswrapper[4815]: I1205 10:38:55.439953 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:55 crc kubenswrapper[4815]: I1205 10:38:55.441674 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:55 crc kubenswrapper[4815]: I1205 10:38:55.507402 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:56 crc kubenswrapper[4815]: I1205 10:38:56.496899 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:56 crc kubenswrapper[4815]: I1205 10:38:56.554423 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rf9tj"] Dec 05 10:38:58 crc kubenswrapper[4815]: I1205 10:38:58.469231 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rf9tj" podUID="824220d8-ccda-4316-a037-b7a4ab90bf28" containerName="registry-server" containerID="cri-o://2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489" gracePeriod=2 Dec 05 10:38:58 crc kubenswrapper[4815]: I1205 10:38:58.968028 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.085608 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-utilities\") pod \"824220d8-ccda-4316-a037-b7a4ab90bf28\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.085732 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2c29\" (UniqueName: \"kubernetes.io/projected/824220d8-ccda-4316-a037-b7a4ab90bf28-kube-api-access-g2c29\") pod \"824220d8-ccda-4316-a037-b7a4ab90bf28\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.085786 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-catalog-content\") pod \"824220d8-ccda-4316-a037-b7a4ab90bf28\" (UID: \"824220d8-ccda-4316-a037-b7a4ab90bf28\") " Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.086293 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-utilities" (OuterVolumeSpecName: "utilities") pod "824220d8-ccda-4316-a037-b7a4ab90bf28" (UID: "824220d8-ccda-4316-a037-b7a4ab90bf28"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.096693 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/824220d8-ccda-4316-a037-b7a4ab90bf28-kube-api-access-g2c29" (OuterVolumeSpecName: "kube-api-access-g2c29") pod "824220d8-ccda-4316-a037-b7a4ab90bf28" (UID: "824220d8-ccda-4316-a037-b7a4ab90bf28"). InnerVolumeSpecName "kube-api-access-g2c29". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.105406 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "824220d8-ccda-4316-a037-b7a4ab90bf28" (UID: "824220d8-ccda-4316-a037-b7a4ab90bf28"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.188860 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2c29\" (UniqueName: \"kubernetes.io/projected/824220d8-ccda-4316-a037-b7a4ab90bf28-kube-api-access-g2c29\") on node \"crc\" DevicePath \"\"" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.189219 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.189462 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/824220d8-ccda-4316-a037-b7a4ab90bf28-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.479990 4815 generic.go:334] "Generic (PLEG): container finished" podID="824220d8-ccda-4316-a037-b7a4ab90bf28" containerID="2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489" exitCode=0 Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.480065 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rf9tj" event={"ID":"824220d8-ccda-4316-a037-b7a4ab90bf28","Type":"ContainerDied","Data":"2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489"} Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.480094 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rf9tj" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.480366 4815 scope.go:117] "RemoveContainer" containerID="2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.480333 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rf9tj" event={"ID":"824220d8-ccda-4316-a037-b7a4ab90bf28","Type":"ContainerDied","Data":"a730ddbce109337283181cab8b327cde8a96f6056dc81c4b48323f94bc6c9863"} Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.503745 4815 scope.go:117] "RemoveContainer" containerID="02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.511603 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rf9tj"] Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.547353 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rf9tj"] Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.555433 4815 scope.go:117] "RemoveContainer" containerID="92520bc899d153147efa8f8d3f7246e0af4e657888f9875845d501cd541b5ab9" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.594055 4815 scope.go:117] "RemoveContainer" containerID="2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489" Dec 05 10:38:59 crc kubenswrapper[4815]: E1205 10:38:59.594914 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489\": container with ID starting with 2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489 not found: ID does not exist" containerID="2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.594960 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489"} err="failed to get container status \"2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489\": rpc error: code = NotFound desc = could not find container \"2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489\": container with ID starting with 2d5877aa0bd586a740067fe519149e137e646bafecd855ecab61706344787489 not found: ID does not exist" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.594988 4815 scope.go:117] "RemoveContainer" containerID="02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83" Dec 05 10:38:59 crc kubenswrapper[4815]: E1205 10:38:59.595268 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83\": container with ID starting with 02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83 not found: ID does not exist" containerID="02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.595298 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83"} err="failed to get container status \"02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83\": rpc error: code = NotFound desc = could not find container \"02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83\": container with ID starting with 02df9b642996facd75f7da8b68cae1d314d16135007dc6dccb2f39d122d74a83 not found: ID does not exist" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.595323 4815 scope.go:117] "RemoveContainer" containerID="92520bc899d153147efa8f8d3f7246e0af4e657888f9875845d501cd541b5ab9" Dec 05 10:38:59 crc kubenswrapper[4815]: E1205 10:38:59.595619 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92520bc899d153147efa8f8d3f7246e0af4e657888f9875845d501cd541b5ab9\": container with ID starting with 92520bc899d153147efa8f8d3f7246e0af4e657888f9875845d501cd541b5ab9 not found: ID does not exist" containerID="92520bc899d153147efa8f8d3f7246e0af4e657888f9875845d501cd541b5ab9" Dec 05 10:38:59 crc kubenswrapper[4815]: I1205 10:38:59.595640 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92520bc899d153147efa8f8d3f7246e0af4e657888f9875845d501cd541b5ab9"} err="failed to get container status \"92520bc899d153147efa8f8d3f7246e0af4e657888f9875845d501cd541b5ab9\": rpc error: code = NotFound desc = could not find container \"92520bc899d153147efa8f8d3f7246e0af4e657888f9875845d501cd541b5ab9\": container with ID starting with 92520bc899d153147efa8f8d3f7246e0af4e657888f9875845d501cd541b5ab9 not found: ID does not exist" Dec 05 10:39:01 crc kubenswrapper[4815]: I1205 10:39:01.430998 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="824220d8-ccda-4316-a037-b7a4ab90bf28" path="/var/lib/kubelet/pods/824220d8-ccda-4316-a037-b7a4ab90bf28/volumes" Dec 05 10:39:09 crc kubenswrapper[4815]: I1205 10:39:09.568385 4815 generic.go:334] "Generic (PLEG): container finished" podID="96c22dce-eb0e-42fd-b689-d50a5e68c8fa" containerID="17ba3448a891bb2c597210fa2bcd29cedf664200ac5c0bad14979df7325f8f23" exitCode=0 Dec 05 10:39:09 crc kubenswrapper[4815]: I1205 10:39:09.568509 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lq64j/crc-debug-9fqqt" event={"ID":"96c22dce-eb0e-42fd-b689-d50a5e68c8fa","Type":"ContainerDied","Data":"17ba3448a891bb2c597210fa2bcd29cedf664200ac5c0bad14979df7325f8f23"} Dec 05 10:39:10 crc kubenswrapper[4815]: I1205 10:39:10.679346 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-9fqqt" Dec 05 10:39:10 crc kubenswrapper[4815]: I1205 10:39:10.716177 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lq64j/crc-debug-9fqqt"] Dec 05 10:39:10 crc kubenswrapper[4815]: I1205 10:39:10.741293 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lq64j/crc-debug-9fqqt"] Dec 05 10:39:10 crc kubenswrapper[4815]: I1205 10:39:10.856481 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnkhd\" (UniqueName: \"kubernetes.io/projected/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-kube-api-access-vnkhd\") pod \"96c22dce-eb0e-42fd-b689-d50a5e68c8fa\" (UID: \"96c22dce-eb0e-42fd-b689-d50a5e68c8fa\") " Dec 05 10:39:10 crc kubenswrapper[4815]: I1205 10:39:10.856573 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-host\") pod \"96c22dce-eb0e-42fd-b689-d50a5e68c8fa\" (UID: \"96c22dce-eb0e-42fd-b689-d50a5e68c8fa\") " Dec 05 10:39:10 crc kubenswrapper[4815]: I1205 10:39:10.857199 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-host" (OuterVolumeSpecName: "host") pod "96c22dce-eb0e-42fd-b689-d50a5e68c8fa" (UID: "96c22dce-eb0e-42fd-b689-d50a5e68c8fa"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:39:10 crc kubenswrapper[4815]: I1205 10:39:10.883661 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-kube-api-access-vnkhd" (OuterVolumeSpecName: "kube-api-access-vnkhd") pod "96c22dce-eb0e-42fd-b689-d50a5e68c8fa" (UID: "96c22dce-eb0e-42fd-b689-d50a5e68c8fa"). InnerVolumeSpecName "kube-api-access-vnkhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:39:10 crc kubenswrapper[4815]: I1205 10:39:10.958355 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnkhd\" (UniqueName: \"kubernetes.io/projected/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-kube-api-access-vnkhd\") on node \"crc\" DevicePath \"\"" Dec 05 10:39:10 crc kubenswrapper[4815]: I1205 10:39:10.958388 4815 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/96c22dce-eb0e-42fd-b689-d50a5e68c8fa-host\") on node \"crc\" DevicePath \"\"" Dec 05 10:39:11 crc kubenswrapper[4815]: I1205 10:39:11.432379 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96c22dce-eb0e-42fd-b689-d50a5e68c8fa" path="/var/lib/kubelet/pods/96c22dce-eb0e-42fd-b689-d50a5e68c8fa/volumes" Dec 05 10:39:11 crc kubenswrapper[4815]: I1205 10:39:11.589102 4815 scope.go:117] "RemoveContainer" containerID="17ba3448a891bb2c597210fa2bcd29cedf664200ac5c0bad14979df7325f8f23" Dec 05 10:39:11 crc kubenswrapper[4815]: I1205 10:39:11.589187 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-9fqqt" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.382883 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lq64j/crc-debug-pq2gw"] Dec 05 10:39:12 crc kubenswrapper[4815]: E1205 10:39:12.383611 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824220d8-ccda-4316-a037-b7a4ab90bf28" containerName="extract-content" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.383646 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="824220d8-ccda-4316-a037-b7a4ab90bf28" containerName="extract-content" Dec 05 10:39:12 crc kubenswrapper[4815]: E1205 10:39:12.383673 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824220d8-ccda-4316-a037-b7a4ab90bf28" containerName="registry-server" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.383679 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="824220d8-ccda-4316-a037-b7a4ab90bf28" containerName="registry-server" Dec 05 10:39:12 crc kubenswrapper[4815]: E1205 10:39:12.383701 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824220d8-ccda-4316-a037-b7a4ab90bf28" containerName="extract-utilities" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.383710 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="824220d8-ccda-4316-a037-b7a4ab90bf28" containerName="extract-utilities" Dec 05 10:39:12 crc kubenswrapper[4815]: E1205 10:39:12.383744 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96c22dce-eb0e-42fd-b689-d50a5e68c8fa" containerName="container-00" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.383751 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="96c22dce-eb0e-42fd-b689-d50a5e68c8fa" containerName="container-00" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.384049 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="96c22dce-eb0e-42fd-b689-d50a5e68c8fa" containerName="container-00" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.384070 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="824220d8-ccda-4316-a037-b7a4ab90bf28" containerName="registry-server" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.387844 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-pq2gw" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.395228 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffvtp\" (UniqueName: \"kubernetes.io/projected/26d2188d-58b5-47a7-a95c-2df5d9b31e98-kube-api-access-ffvtp\") pod \"crc-debug-pq2gw\" (UID: \"26d2188d-58b5-47a7-a95c-2df5d9b31e98\") " pod="openshift-must-gather-lq64j/crc-debug-pq2gw" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.395921 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/26d2188d-58b5-47a7-a95c-2df5d9b31e98-host\") pod \"crc-debug-pq2gw\" (UID: \"26d2188d-58b5-47a7-a95c-2df5d9b31e98\") " pod="openshift-must-gather-lq64j/crc-debug-pq2gw" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.497483 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/26d2188d-58b5-47a7-a95c-2df5d9b31e98-host\") pod \"crc-debug-pq2gw\" (UID: \"26d2188d-58b5-47a7-a95c-2df5d9b31e98\") " pod="openshift-must-gather-lq64j/crc-debug-pq2gw" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.497635 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffvtp\" (UniqueName: \"kubernetes.io/projected/26d2188d-58b5-47a7-a95c-2df5d9b31e98-kube-api-access-ffvtp\") pod \"crc-debug-pq2gw\" (UID: \"26d2188d-58b5-47a7-a95c-2df5d9b31e98\") " pod="openshift-must-gather-lq64j/crc-debug-pq2gw" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.498348 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/26d2188d-58b5-47a7-a95c-2df5d9b31e98-host\") pod \"crc-debug-pq2gw\" (UID: \"26d2188d-58b5-47a7-a95c-2df5d9b31e98\") " pod="openshift-must-gather-lq64j/crc-debug-pq2gw" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.517002 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffvtp\" (UniqueName: \"kubernetes.io/projected/26d2188d-58b5-47a7-a95c-2df5d9b31e98-kube-api-access-ffvtp\") pod \"crc-debug-pq2gw\" (UID: \"26d2188d-58b5-47a7-a95c-2df5d9b31e98\") " pod="openshift-must-gather-lq64j/crc-debug-pq2gw" Dec 05 10:39:12 crc kubenswrapper[4815]: I1205 10:39:12.711328 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-pq2gw" Dec 05 10:39:13 crc kubenswrapper[4815]: I1205 10:39:13.607907 4815 generic.go:334] "Generic (PLEG): container finished" podID="26d2188d-58b5-47a7-a95c-2df5d9b31e98" containerID="7e99bc718e2347927ca3b445af0c3d1c50c27152c414718e7830642bd2b149e4" exitCode=0 Dec 05 10:39:13 crc kubenswrapper[4815]: I1205 10:39:13.607990 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lq64j/crc-debug-pq2gw" event={"ID":"26d2188d-58b5-47a7-a95c-2df5d9b31e98","Type":"ContainerDied","Data":"7e99bc718e2347927ca3b445af0c3d1c50c27152c414718e7830642bd2b149e4"} Dec 05 10:39:13 crc kubenswrapper[4815]: I1205 10:39:13.608162 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lq64j/crc-debug-pq2gw" event={"ID":"26d2188d-58b5-47a7-a95c-2df5d9b31e98","Type":"ContainerStarted","Data":"42c536de004ed6f0318881ffdd1230ad0cb5d014663d7cd4206bbe45d4fdfb85"} Dec 05 10:39:13 crc kubenswrapper[4815]: I1205 10:39:13.985437 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lq64j/crc-debug-pq2gw"] Dec 05 10:39:13 crc kubenswrapper[4815]: I1205 10:39:13.997695 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lq64j/crc-debug-pq2gw"] Dec 05 10:39:14 crc kubenswrapper[4815]: I1205 10:39:14.724734 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-pq2gw" Dec 05 10:39:14 crc kubenswrapper[4815]: I1205 10:39:14.791010 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/26d2188d-58b5-47a7-a95c-2df5d9b31e98-host\") pod \"26d2188d-58b5-47a7-a95c-2df5d9b31e98\" (UID: \"26d2188d-58b5-47a7-a95c-2df5d9b31e98\") " Dec 05 10:39:14 crc kubenswrapper[4815]: I1205 10:39:14.791138 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26d2188d-58b5-47a7-a95c-2df5d9b31e98-host" (OuterVolumeSpecName: "host") pod "26d2188d-58b5-47a7-a95c-2df5d9b31e98" (UID: "26d2188d-58b5-47a7-a95c-2df5d9b31e98"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:39:14 crc kubenswrapper[4815]: I1205 10:39:14.791250 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffvtp\" (UniqueName: \"kubernetes.io/projected/26d2188d-58b5-47a7-a95c-2df5d9b31e98-kube-api-access-ffvtp\") pod \"26d2188d-58b5-47a7-a95c-2df5d9b31e98\" (UID: \"26d2188d-58b5-47a7-a95c-2df5d9b31e98\") " Dec 05 10:39:14 crc kubenswrapper[4815]: I1205 10:39:14.791621 4815 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/26d2188d-58b5-47a7-a95c-2df5d9b31e98-host\") on node \"crc\" DevicePath \"\"" Dec 05 10:39:14 crc kubenswrapper[4815]: I1205 10:39:14.798935 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26d2188d-58b5-47a7-a95c-2df5d9b31e98-kube-api-access-ffvtp" (OuterVolumeSpecName: "kube-api-access-ffvtp") pod "26d2188d-58b5-47a7-a95c-2df5d9b31e98" (UID: "26d2188d-58b5-47a7-a95c-2df5d9b31e98"). InnerVolumeSpecName "kube-api-access-ffvtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:39:14 crc kubenswrapper[4815]: I1205 10:39:14.893235 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffvtp\" (UniqueName: \"kubernetes.io/projected/26d2188d-58b5-47a7-a95c-2df5d9b31e98-kube-api-access-ffvtp\") on node \"crc\" DevicePath \"\"" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.310612 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-lq64j/crc-debug-nb9qn"] Dec 05 10:39:15 crc kubenswrapper[4815]: E1205 10:39:15.311165 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26d2188d-58b5-47a7-a95c-2df5d9b31e98" containerName="container-00" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.311193 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="26d2188d-58b5-47a7-a95c-2df5d9b31e98" containerName="container-00" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.311455 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="26d2188d-58b5-47a7-a95c-2df5d9b31e98" containerName="container-00" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.312339 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-nb9qn" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.423216 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plzls\" (UniqueName: \"kubernetes.io/projected/462616a5-196b-450f-aa12-ab7ea1917f3c-kube-api-access-plzls\") pod \"crc-debug-nb9qn\" (UID: \"462616a5-196b-450f-aa12-ab7ea1917f3c\") " pod="openshift-must-gather-lq64j/crc-debug-nb9qn" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.423762 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/462616a5-196b-450f-aa12-ab7ea1917f3c-host\") pod \"crc-debug-nb9qn\" (UID: \"462616a5-196b-450f-aa12-ab7ea1917f3c\") " pod="openshift-must-gather-lq64j/crc-debug-nb9qn" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.432142 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26d2188d-58b5-47a7-a95c-2df5d9b31e98" path="/var/lib/kubelet/pods/26d2188d-58b5-47a7-a95c-2df5d9b31e98/volumes" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.525980 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plzls\" (UniqueName: \"kubernetes.io/projected/462616a5-196b-450f-aa12-ab7ea1917f3c-kube-api-access-plzls\") pod \"crc-debug-nb9qn\" (UID: \"462616a5-196b-450f-aa12-ab7ea1917f3c\") " pod="openshift-must-gather-lq64j/crc-debug-nb9qn" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.526162 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/462616a5-196b-450f-aa12-ab7ea1917f3c-host\") pod \"crc-debug-nb9qn\" (UID: \"462616a5-196b-450f-aa12-ab7ea1917f3c\") " pod="openshift-must-gather-lq64j/crc-debug-nb9qn" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.526334 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/462616a5-196b-450f-aa12-ab7ea1917f3c-host\") pod \"crc-debug-nb9qn\" (UID: \"462616a5-196b-450f-aa12-ab7ea1917f3c\") " pod="openshift-must-gather-lq64j/crc-debug-nb9qn" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.544679 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plzls\" (UniqueName: \"kubernetes.io/projected/462616a5-196b-450f-aa12-ab7ea1917f3c-kube-api-access-plzls\") pod \"crc-debug-nb9qn\" (UID: \"462616a5-196b-450f-aa12-ab7ea1917f3c\") " pod="openshift-must-gather-lq64j/crc-debug-nb9qn" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.627743 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-nb9qn" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.629788 4815 scope.go:117] "RemoveContainer" containerID="7e99bc718e2347927ca3b445af0c3d1c50c27152c414718e7830642bd2b149e4" Dec 05 10:39:15 crc kubenswrapper[4815]: I1205 10:39:15.629850 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-pq2gw" Dec 05 10:39:15 crc kubenswrapper[4815]: W1205 10:39:15.708860 4815 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod462616a5_196b_450f_aa12_ab7ea1917f3c.slice/crio-4f1a7df82799e5cd505a854784b48cb6cc951c0359a09d2741e987e1a7170993 WatchSource:0}: Error finding container 4f1a7df82799e5cd505a854784b48cb6cc951c0359a09d2741e987e1a7170993: Status 404 returned error can't find the container with id 4f1a7df82799e5cd505a854784b48cb6cc951c0359a09d2741e987e1a7170993 Dec 05 10:39:16 crc kubenswrapper[4815]: I1205 10:39:16.643547 4815 generic.go:334] "Generic (PLEG): container finished" podID="462616a5-196b-450f-aa12-ab7ea1917f3c" containerID="a2a2e5b95a11df9093a8b9981d4cbd4316ecf8216e7362f50b9c0e118ac11b45" exitCode=0 Dec 05 10:39:16 crc kubenswrapper[4815]: I1205 10:39:16.643632 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lq64j/crc-debug-nb9qn" event={"ID":"462616a5-196b-450f-aa12-ab7ea1917f3c","Type":"ContainerDied","Data":"a2a2e5b95a11df9093a8b9981d4cbd4316ecf8216e7362f50b9c0e118ac11b45"} Dec 05 10:39:16 crc kubenswrapper[4815]: I1205 10:39:16.643903 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lq64j/crc-debug-nb9qn" event={"ID":"462616a5-196b-450f-aa12-ab7ea1917f3c","Type":"ContainerStarted","Data":"4f1a7df82799e5cd505a854784b48cb6cc951c0359a09d2741e987e1a7170993"} Dec 05 10:39:16 crc kubenswrapper[4815]: I1205 10:39:16.696323 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lq64j/crc-debug-nb9qn"] Dec 05 10:39:16 crc kubenswrapper[4815]: I1205 10:39:16.710798 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lq64j/crc-debug-nb9qn"] Dec 05 10:39:17 crc kubenswrapper[4815]: I1205 10:39:17.759668 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-nb9qn" Dec 05 10:39:17 crc kubenswrapper[4815]: I1205 10:39:17.903859 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/462616a5-196b-450f-aa12-ab7ea1917f3c-host\") pod \"462616a5-196b-450f-aa12-ab7ea1917f3c\" (UID: \"462616a5-196b-450f-aa12-ab7ea1917f3c\") " Dec 05 10:39:17 crc kubenswrapper[4815]: I1205 10:39:17.903964 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plzls\" (UniqueName: \"kubernetes.io/projected/462616a5-196b-450f-aa12-ab7ea1917f3c-kube-api-access-plzls\") pod \"462616a5-196b-450f-aa12-ab7ea1917f3c\" (UID: \"462616a5-196b-450f-aa12-ab7ea1917f3c\") " Dec 05 10:39:17 crc kubenswrapper[4815]: I1205 10:39:17.905651 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/462616a5-196b-450f-aa12-ab7ea1917f3c-host" (OuterVolumeSpecName: "host") pod "462616a5-196b-450f-aa12-ab7ea1917f3c" (UID: "462616a5-196b-450f-aa12-ab7ea1917f3c"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 05 10:39:17 crc kubenswrapper[4815]: I1205 10:39:17.944798 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/462616a5-196b-450f-aa12-ab7ea1917f3c-kube-api-access-plzls" (OuterVolumeSpecName: "kube-api-access-plzls") pod "462616a5-196b-450f-aa12-ab7ea1917f3c" (UID: "462616a5-196b-450f-aa12-ab7ea1917f3c"). InnerVolumeSpecName "kube-api-access-plzls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:39:18 crc kubenswrapper[4815]: I1205 10:39:18.007273 4815 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/462616a5-196b-450f-aa12-ab7ea1917f3c-host\") on node \"crc\" DevicePath \"\"" Dec 05 10:39:18 crc kubenswrapper[4815]: I1205 10:39:18.007532 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plzls\" (UniqueName: \"kubernetes.io/projected/462616a5-196b-450f-aa12-ab7ea1917f3c-kube-api-access-plzls\") on node \"crc\" DevicePath \"\"" Dec 05 10:39:18 crc kubenswrapper[4815]: I1205 10:39:18.674622 4815 scope.go:117] "RemoveContainer" containerID="a2a2e5b95a11df9093a8b9981d4cbd4316ecf8216e7362f50b9c0e118ac11b45" Dec 05 10:39:18 crc kubenswrapper[4815]: I1205 10:39:18.674777 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/crc-debug-nb9qn" Dec 05 10:39:19 crc kubenswrapper[4815]: I1205 10:39:19.439024 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="462616a5-196b-450f-aa12-ab7ea1917f3c" path="/var/lib/kubelet/pods/462616a5-196b-450f-aa12-ab7ea1917f3c/volumes" Dec 05 10:39:20 crc kubenswrapper[4815]: I1205 10:39:20.192184 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:39:20 crc kubenswrapper[4815]: I1205 10:39:20.192244 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.032555 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-spxzj"] Dec 05 10:39:26 crc kubenswrapper[4815]: E1205 10:39:26.033463 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="462616a5-196b-450f-aa12-ab7ea1917f3c" containerName="container-00" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.033498 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="462616a5-196b-450f-aa12-ab7ea1917f3c" containerName="container-00" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.033733 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="462616a5-196b-450f-aa12-ab7ea1917f3c" containerName="container-00" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.035169 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.047591 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-spxzj"] Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.165876 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvx2v\" (UniqueName: \"kubernetes.io/projected/e5a18c79-e0ba-423c-8feb-889118e6f81c-kube-api-access-pvx2v\") pod \"redhat-operators-spxzj\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.165941 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-catalog-content\") pod \"redhat-operators-spxzj\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.166129 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-utilities\") pod \"redhat-operators-spxzj\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.268001 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-utilities\") pod \"redhat-operators-spxzj\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.268148 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvx2v\" (UniqueName: \"kubernetes.io/projected/e5a18c79-e0ba-423c-8feb-889118e6f81c-kube-api-access-pvx2v\") pod \"redhat-operators-spxzj\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.268178 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-catalog-content\") pod \"redhat-operators-spxzj\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.268630 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-utilities\") pod \"redhat-operators-spxzj\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.268742 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-catalog-content\") pod \"redhat-operators-spxzj\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.291602 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvx2v\" (UniqueName: \"kubernetes.io/projected/e5a18c79-e0ba-423c-8feb-889118e6f81c-kube-api-access-pvx2v\") pod \"redhat-operators-spxzj\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.358607 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:26 crc kubenswrapper[4815]: I1205 10:39:26.983889 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-spxzj"] Dec 05 10:39:27 crc kubenswrapper[4815]: I1205 10:39:27.756713 4815 generic.go:334] "Generic (PLEG): container finished" podID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerID="eedfd983d127bf756210cb8d84ae176b0fc69fb07b5607aa322fcafedb670adb" exitCode=0 Dec 05 10:39:27 crc kubenswrapper[4815]: I1205 10:39:27.756834 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spxzj" event={"ID":"e5a18c79-e0ba-423c-8feb-889118e6f81c","Type":"ContainerDied","Data":"eedfd983d127bf756210cb8d84ae176b0fc69fb07b5607aa322fcafedb670adb"} Dec 05 10:39:27 crc kubenswrapper[4815]: I1205 10:39:27.757015 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spxzj" event={"ID":"e5a18c79-e0ba-423c-8feb-889118e6f81c","Type":"ContainerStarted","Data":"1123ebe007e7976e8a0d5c2d9aaeca154db3d5b178cdb8e0ed6204077068c5ac"} Dec 05 10:39:28 crc kubenswrapper[4815]: I1205 10:39:28.766570 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spxzj" event={"ID":"e5a18c79-e0ba-423c-8feb-889118e6f81c","Type":"ContainerStarted","Data":"d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2"} Dec 05 10:39:33 crc kubenswrapper[4815]: I1205 10:39:33.089478 4815 generic.go:334] "Generic (PLEG): container finished" podID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerID="d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2" exitCode=0 Dec 05 10:39:33 crc kubenswrapper[4815]: I1205 10:39:33.089581 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spxzj" event={"ID":"e5a18c79-e0ba-423c-8feb-889118e6f81c","Type":"ContainerDied","Data":"d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2"} Dec 05 10:39:34 crc kubenswrapper[4815]: I1205 10:39:34.099303 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spxzj" event={"ID":"e5a18c79-e0ba-423c-8feb-889118e6f81c","Type":"ContainerStarted","Data":"03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae"} Dec 05 10:39:34 crc kubenswrapper[4815]: I1205 10:39:34.130576 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-spxzj" podStartSLOduration=2.328989783 podStartE2EDuration="8.130544796s" podCreationTimestamp="2025-12-05 10:39:26 +0000 UTC" firstStartedPulling="2025-12-05 10:39:27.759652821 +0000 UTC m=+5566.638259658" lastFinishedPulling="2025-12-05 10:39:33.561207834 +0000 UTC m=+5572.439814671" observedRunningTime="2025-12-05 10:39:34.120412316 +0000 UTC m=+5572.999019153" watchObservedRunningTime="2025-12-05 10:39:34.130544796 +0000 UTC m=+5573.009151633" Dec 05 10:39:36 crc kubenswrapper[4815]: I1205 10:39:36.358784 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:36 crc kubenswrapper[4815]: I1205 10:39:36.359988 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:37 crc kubenswrapper[4815]: I1205 10:39:37.426432 4815 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-spxzj" podUID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerName="registry-server" probeResult="failure" output=< Dec 05 10:39:37 crc kubenswrapper[4815]: timeout: failed to connect service ":50051" within 1s Dec 05 10:39:37 crc kubenswrapper[4815]: > Dec 05 10:39:46 crc kubenswrapper[4815]: I1205 10:39:46.407351 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:46 crc kubenswrapper[4815]: I1205 10:39:46.458956 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:46 crc kubenswrapper[4815]: I1205 10:39:46.646675 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-spxzj"] Dec 05 10:39:48 crc kubenswrapper[4815]: I1205 10:39:48.225014 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-spxzj" podUID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerName="registry-server" containerID="cri-o://03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae" gracePeriod=2 Dec 05 10:39:48 crc kubenswrapper[4815]: I1205 10:39:48.783560 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:48 crc kubenswrapper[4815]: I1205 10:39:48.849469 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvx2v\" (UniqueName: \"kubernetes.io/projected/e5a18c79-e0ba-423c-8feb-889118e6f81c-kube-api-access-pvx2v\") pod \"e5a18c79-e0ba-423c-8feb-889118e6f81c\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " Dec 05 10:39:48 crc kubenswrapper[4815]: I1205 10:39:48.849567 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-catalog-content\") pod \"e5a18c79-e0ba-423c-8feb-889118e6f81c\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " Dec 05 10:39:48 crc kubenswrapper[4815]: I1205 10:39:48.849625 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-utilities\") pod \"e5a18c79-e0ba-423c-8feb-889118e6f81c\" (UID: \"e5a18c79-e0ba-423c-8feb-889118e6f81c\") " Dec 05 10:39:48 crc kubenswrapper[4815]: I1205 10:39:48.851802 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-utilities" (OuterVolumeSpecName: "utilities") pod "e5a18c79-e0ba-423c-8feb-889118e6f81c" (UID: "e5a18c79-e0ba-423c-8feb-889118e6f81c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:39:48 crc kubenswrapper[4815]: I1205 10:39:48.874224 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5a18c79-e0ba-423c-8feb-889118e6f81c-kube-api-access-pvx2v" (OuterVolumeSpecName: "kube-api-access-pvx2v") pod "e5a18c79-e0ba-423c-8feb-889118e6f81c" (UID: "e5a18c79-e0ba-423c-8feb-889118e6f81c"). InnerVolumeSpecName "kube-api-access-pvx2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:39:48 crc kubenswrapper[4815]: I1205 10:39:48.952922 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvx2v\" (UniqueName: \"kubernetes.io/projected/e5a18c79-e0ba-423c-8feb-889118e6f81c-kube-api-access-pvx2v\") on node \"crc\" DevicePath \"\"" Dec 05 10:39:48 crc kubenswrapper[4815]: I1205 10:39:48.952959 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:39:48 crc kubenswrapper[4815]: I1205 10:39:48.974575 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5a18c79-e0ba-423c-8feb-889118e6f81c" (UID: "e5a18c79-e0ba-423c-8feb-889118e6f81c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.054348 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5a18c79-e0ba-423c-8feb-889118e6f81c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.236089 4815 generic.go:334] "Generic (PLEG): container finished" podID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerID="03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae" exitCode=0 Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.236134 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spxzj" event={"ID":"e5a18c79-e0ba-423c-8feb-889118e6f81c","Type":"ContainerDied","Data":"03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae"} Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.236165 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-spxzj" event={"ID":"e5a18c79-e0ba-423c-8feb-889118e6f81c","Type":"ContainerDied","Data":"1123ebe007e7976e8a0d5c2d9aaeca154db3d5b178cdb8e0ed6204077068c5ac"} Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.236182 4815 scope.go:117] "RemoveContainer" containerID="03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.236202 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-spxzj" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.262762 4815 scope.go:117] "RemoveContainer" containerID="d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.283066 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-spxzj"] Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.294377 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-spxzj"] Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.295146 4815 scope.go:117] "RemoveContainer" containerID="eedfd983d127bf756210cb8d84ae176b0fc69fb07b5607aa322fcafedb670adb" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.369230 4815 scope.go:117] "RemoveContainer" containerID="03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae" Dec 05 10:39:49 crc kubenswrapper[4815]: E1205 10:39:49.369816 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae\": container with ID starting with 03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae not found: ID does not exist" containerID="03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.369848 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae"} err="failed to get container status \"03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae\": rpc error: code = NotFound desc = could not find container \"03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae\": container with ID starting with 03d1a998ac6908d55ee2cd2e4b8456b0e7869bf2a37396739ee259bfd57ffeae not found: ID does not exist" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.369870 4815 scope.go:117] "RemoveContainer" containerID="d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2" Dec 05 10:39:49 crc kubenswrapper[4815]: E1205 10:39:49.370149 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2\": container with ID starting with d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2 not found: ID does not exist" containerID="d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.370173 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2"} err="failed to get container status \"d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2\": rpc error: code = NotFound desc = could not find container \"d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2\": container with ID starting with d2c9b965e9e60614cf4d3910160b6392d0c5538f15a45199a09754b1f04540a2 not found: ID does not exist" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.370187 4815 scope.go:117] "RemoveContainer" containerID="eedfd983d127bf756210cb8d84ae176b0fc69fb07b5607aa322fcafedb670adb" Dec 05 10:39:49 crc kubenswrapper[4815]: E1205 10:39:49.370450 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eedfd983d127bf756210cb8d84ae176b0fc69fb07b5607aa322fcafedb670adb\": container with ID starting with eedfd983d127bf756210cb8d84ae176b0fc69fb07b5607aa322fcafedb670adb not found: ID does not exist" containerID="eedfd983d127bf756210cb8d84ae176b0fc69fb07b5607aa322fcafedb670adb" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.370472 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eedfd983d127bf756210cb8d84ae176b0fc69fb07b5607aa322fcafedb670adb"} err="failed to get container status \"eedfd983d127bf756210cb8d84ae176b0fc69fb07b5607aa322fcafedb670adb\": rpc error: code = NotFound desc = could not find container \"eedfd983d127bf756210cb8d84ae176b0fc69fb07b5607aa322fcafedb670adb\": container with ID starting with eedfd983d127bf756210cb8d84ae176b0fc69fb07b5607aa322fcafedb670adb not found: ID does not exist" Dec 05 10:39:49 crc kubenswrapper[4815]: I1205 10:39:49.463300 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5a18c79-e0ba-423c-8feb-889118e6f81c" path="/var/lib/kubelet/pods/e5a18c79-e0ba-423c-8feb-889118e6f81c/volumes" Dec 05 10:39:50 crc kubenswrapper[4815]: I1205 10:39:50.192737 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:39:50 crc kubenswrapper[4815]: I1205 10:39:50.192827 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.441076 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-spcgb"] Dec 05 10:39:59 crc kubenswrapper[4815]: E1205 10:39:59.443295 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerName="extract-utilities" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.443450 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerName="extract-utilities" Dec 05 10:39:59 crc kubenswrapper[4815]: E1205 10:39:59.443600 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerName="registry-server" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.443705 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerName="registry-server" Dec 05 10:39:59 crc kubenswrapper[4815]: E1205 10:39:59.443811 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerName="extract-content" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.443921 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerName="extract-content" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.444272 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5a18c79-e0ba-423c-8feb-889118e6f81c" containerName="registry-server" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.446210 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-spcgb"] Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.446873 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.579034 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-catalog-content\") pod \"certified-operators-spcgb\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.579163 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf9rn\" (UniqueName: \"kubernetes.io/projected/56156298-bd80-41f4-bbe7-6c0603005afd-kube-api-access-sf9rn\") pod \"certified-operators-spcgb\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.579260 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-utilities\") pod \"certified-operators-spcgb\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.680938 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-utilities\") pod \"certified-operators-spcgb\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.681020 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-catalog-content\") pod \"certified-operators-spcgb\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.681138 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf9rn\" (UniqueName: \"kubernetes.io/projected/56156298-bd80-41f4-bbe7-6c0603005afd-kube-api-access-sf9rn\") pod \"certified-operators-spcgb\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.681412 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-utilities\") pod \"certified-operators-spcgb\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.681602 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-catalog-content\") pod \"certified-operators-spcgb\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.700076 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf9rn\" (UniqueName: \"kubernetes.io/projected/56156298-bd80-41f4-bbe7-6c0603005afd-kube-api-access-sf9rn\") pod \"certified-operators-spcgb\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:39:59 crc kubenswrapper[4815]: I1205 10:39:59.776328 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:40:00 crc kubenswrapper[4815]: I1205 10:40:00.394851 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-spcgb"] Dec 05 10:40:01 crc kubenswrapper[4815]: I1205 10:40:01.371345 4815 generic.go:334] "Generic (PLEG): container finished" podID="56156298-bd80-41f4-bbe7-6c0603005afd" containerID="84cea7a465cde9bda650112589107dcfef3e0f79df988558dca9ae3cc0de4f62" exitCode=0 Dec 05 10:40:01 crc kubenswrapper[4815]: I1205 10:40:01.371397 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-spcgb" event={"ID":"56156298-bd80-41f4-bbe7-6c0603005afd","Type":"ContainerDied","Data":"84cea7a465cde9bda650112589107dcfef3e0f79df988558dca9ae3cc0de4f62"} Dec 05 10:40:01 crc kubenswrapper[4815]: I1205 10:40:01.371892 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-spcgb" event={"ID":"56156298-bd80-41f4-bbe7-6c0603005afd","Type":"ContainerStarted","Data":"38f3e63ab04bd3c9672e3bbee341043206aedd27ebfadcd399e468dc93c4cb9d"} Dec 05 10:40:03 crc kubenswrapper[4815]: I1205 10:40:03.394869 4815 generic.go:334] "Generic (PLEG): container finished" podID="56156298-bd80-41f4-bbe7-6c0603005afd" containerID="c7ede3f1d58e6a12ce5f71d3242ea6b4e3ae50365d28c3ce28ff502f5930e5d6" exitCode=0 Dec 05 10:40:03 crc kubenswrapper[4815]: I1205 10:40:03.394982 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-spcgb" event={"ID":"56156298-bd80-41f4-bbe7-6c0603005afd","Type":"ContainerDied","Data":"c7ede3f1d58e6a12ce5f71d3242ea6b4e3ae50365d28c3ce28ff502f5930e5d6"} Dec 05 10:40:04 crc kubenswrapper[4815]: I1205 10:40:04.408646 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-spcgb" event={"ID":"56156298-bd80-41f4-bbe7-6c0603005afd","Type":"ContainerStarted","Data":"5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8"} Dec 05 10:40:04 crc kubenswrapper[4815]: I1205 10:40:04.435864 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-spcgb" podStartSLOduration=2.993787293 podStartE2EDuration="5.435845921s" podCreationTimestamp="2025-12-05 10:39:59 +0000 UTC" firstStartedPulling="2025-12-05 10:40:01.374650039 +0000 UTC m=+5600.253256886" lastFinishedPulling="2025-12-05 10:40:03.816708657 +0000 UTC m=+5602.695315514" observedRunningTime="2025-12-05 10:40:04.428366755 +0000 UTC m=+5603.306973612" watchObservedRunningTime="2025-12-05 10:40:04.435845921 +0000 UTC m=+5603.314452758" Dec 05 10:40:09 crc kubenswrapper[4815]: I1205 10:40:09.777195 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:40:09 crc kubenswrapper[4815]: I1205 10:40:09.778930 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:40:09 crc kubenswrapper[4815]: I1205 10:40:09.840420 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:40:10 crc kubenswrapper[4815]: I1205 10:40:10.526176 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:40:10 crc kubenswrapper[4815]: I1205 10:40:10.578200 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-spcgb"] Dec 05 10:40:12 crc kubenswrapper[4815]: I1205 10:40:12.502771 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-spcgb" podUID="56156298-bd80-41f4-bbe7-6c0603005afd" containerName="registry-server" containerID="cri-o://5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8" gracePeriod=2 Dec 05 10:40:12 crc kubenswrapper[4815]: I1205 10:40:12.997852 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.142194 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-utilities\") pod \"56156298-bd80-41f4-bbe7-6c0603005afd\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.142282 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf9rn\" (UniqueName: \"kubernetes.io/projected/56156298-bd80-41f4-bbe7-6c0603005afd-kube-api-access-sf9rn\") pod \"56156298-bd80-41f4-bbe7-6c0603005afd\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.142440 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-catalog-content\") pod \"56156298-bd80-41f4-bbe7-6c0603005afd\" (UID: \"56156298-bd80-41f4-bbe7-6c0603005afd\") " Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.143693 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-utilities" (OuterVolumeSpecName: "utilities") pod "56156298-bd80-41f4-bbe7-6c0603005afd" (UID: "56156298-bd80-41f4-bbe7-6c0603005afd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.150732 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56156298-bd80-41f4-bbe7-6c0603005afd-kube-api-access-sf9rn" (OuterVolumeSpecName: "kube-api-access-sf9rn") pod "56156298-bd80-41f4-bbe7-6c0603005afd" (UID: "56156298-bd80-41f4-bbe7-6c0603005afd"). InnerVolumeSpecName "kube-api-access-sf9rn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.225935 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56156298-bd80-41f4-bbe7-6c0603005afd" (UID: "56156298-bd80-41f4-bbe7-6c0603005afd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.244752 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.244797 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf9rn\" (UniqueName: \"kubernetes.io/projected/56156298-bd80-41f4-bbe7-6c0603005afd-kube-api-access-sf9rn\") on node \"crc\" DevicePath \"\"" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.244812 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56156298-bd80-41f4-bbe7-6c0603005afd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.516292 4815 generic.go:334] "Generic (PLEG): container finished" podID="56156298-bd80-41f4-bbe7-6c0603005afd" containerID="5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8" exitCode=0 Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.516351 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-spcgb" event={"ID":"56156298-bd80-41f4-bbe7-6c0603005afd","Type":"ContainerDied","Data":"5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8"} Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.516704 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-spcgb" event={"ID":"56156298-bd80-41f4-bbe7-6c0603005afd","Type":"ContainerDied","Data":"38f3e63ab04bd3c9672e3bbee341043206aedd27ebfadcd399e468dc93c4cb9d"} Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.516724 4815 scope.go:117] "RemoveContainer" containerID="5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.516425 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-spcgb" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.547829 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-spcgb"] Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.556056 4815 scope.go:117] "RemoveContainer" containerID="c7ede3f1d58e6a12ce5f71d3242ea6b4e3ae50365d28c3ce28ff502f5930e5d6" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.558536 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-spcgb"] Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.587629 4815 scope.go:117] "RemoveContainer" containerID="84cea7a465cde9bda650112589107dcfef3e0f79df988558dca9ae3cc0de4f62" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.626226 4815 scope.go:117] "RemoveContainer" containerID="5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8" Dec 05 10:40:13 crc kubenswrapper[4815]: E1205 10:40:13.627250 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8\": container with ID starting with 5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8 not found: ID does not exist" containerID="5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.627283 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8"} err="failed to get container status \"5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8\": rpc error: code = NotFound desc = could not find container \"5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8\": container with ID starting with 5247c83e64dc9f726c175c9f3c402712a794b8c141a16d222c9a341c0fb82fa8 not found: ID does not exist" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.627305 4815 scope.go:117] "RemoveContainer" containerID="c7ede3f1d58e6a12ce5f71d3242ea6b4e3ae50365d28c3ce28ff502f5930e5d6" Dec 05 10:40:13 crc kubenswrapper[4815]: E1205 10:40:13.627746 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7ede3f1d58e6a12ce5f71d3242ea6b4e3ae50365d28c3ce28ff502f5930e5d6\": container with ID starting with c7ede3f1d58e6a12ce5f71d3242ea6b4e3ae50365d28c3ce28ff502f5930e5d6 not found: ID does not exist" containerID="c7ede3f1d58e6a12ce5f71d3242ea6b4e3ae50365d28c3ce28ff502f5930e5d6" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.627766 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7ede3f1d58e6a12ce5f71d3242ea6b4e3ae50365d28c3ce28ff502f5930e5d6"} err="failed to get container status \"c7ede3f1d58e6a12ce5f71d3242ea6b4e3ae50365d28c3ce28ff502f5930e5d6\": rpc error: code = NotFound desc = could not find container \"c7ede3f1d58e6a12ce5f71d3242ea6b4e3ae50365d28c3ce28ff502f5930e5d6\": container with ID starting with c7ede3f1d58e6a12ce5f71d3242ea6b4e3ae50365d28c3ce28ff502f5930e5d6 not found: ID does not exist" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.627779 4815 scope.go:117] "RemoveContainer" containerID="84cea7a465cde9bda650112589107dcfef3e0f79df988558dca9ae3cc0de4f62" Dec 05 10:40:13 crc kubenswrapper[4815]: E1205 10:40:13.628246 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84cea7a465cde9bda650112589107dcfef3e0f79df988558dca9ae3cc0de4f62\": container with ID starting with 84cea7a465cde9bda650112589107dcfef3e0f79df988558dca9ae3cc0de4f62 not found: ID does not exist" containerID="84cea7a465cde9bda650112589107dcfef3e0f79df988558dca9ae3cc0de4f62" Dec 05 10:40:13 crc kubenswrapper[4815]: I1205 10:40:13.628269 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84cea7a465cde9bda650112589107dcfef3e0f79df988558dca9ae3cc0de4f62"} err="failed to get container status \"84cea7a465cde9bda650112589107dcfef3e0f79df988558dca9ae3cc0de4f62\": rpc error: code = NotFound desc = could not find container \"84cea7a465cde9bda650112589107dcfef3e0f79df988558dca9ae3cc0de4f62\": container with ID starting with 84cea7a465cde9bda650112589107dcfef3e0f79df988558dca9ae3cc0de4f62 not found: ID does not exist" Dec 05 10:40:15 crc kubenswrapper[4815]: I1205 10:40:15.430609 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56156298-bd80-41f4-bbe7-6c0603005afd" path="/var/lib/kubelet/pods/56156298-bd80-41f4-bbe7-6c0603005afd/volumes" Dec 05 10:40:20 crc kubenswrapper[4815]: I1205 10:40:20.192630 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:40:20 crc kubenswrapper[4815]: I1205 10:40:20.193472 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:40:20 crc kubenswrapper[4815]: I1205 10:40:20.193565 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 10:40:20 crc kubenswrapper[4815]: I1205 10:40:20.194535 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"36532cd36fe5a7fe6381b37918506e1e1d6db7ef1721fc2254713e8eb9050c93"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:40:20 crc kubenswrapper[4815]: I1205 10:40:20.194614 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://36532cd36fe5a7fe6381b37918506e1e1d6db7ef1721fc2254713e8eb9050c93" gracePeriod=600 Dec 05 10:40:20 crc kubenswrapper[4815]: I1205 10:40:20.600628 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"36532cd36fe5a7fe6381b37918506e1e1d6db7ef1721fc2254713e8eb9050c93"} Dec 05 10:40:20 crc kubenswrapper[4815]: I1205 10:40:20.600962 4815 scope.go:117] "RemoveContainer" containerID="c6ca5c824a8a1ad50ea39f1934ef57c094d216fc522dfb5b79c29acbac01e26f" Dec 05 10:40:20 crc kubenswrapper[4815]: I1205 10:40:20.600571 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="36532cd36fe5a7fe6381b37918506e1e1d6db7ef1721fc2254713e8eb9050c93" exitCode=0 Dec 05 10:40:21 crc kubenswrapper[4815]: I1205 10:40:21.616251 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94"} Dec 05 10:40:42 crc kubenswrapper[4815]: I1205 10:40:42.772394 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-588897d4d8-8nmk9_d7266a34-7fa1-442b-9e83-e66665a5cd12/barbican-api/0.log" Dec 05 10:40:43 crc kubenswrapper[4815]: I1205 10:40:43.081823 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-588897d4d8-8nmk9_d7266a34-7fa1-442b-9e83-e66665a5cd12/barbican-api-log/0.log" Dec 05 10:40:43 crc kubenswrapper[4815]: I1205 10:40:43.141653 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-596686cb9d-rwjpw_3f638151-76e9-429e-b283-d741ad857276/barbican-keystone-listener/0.log" Dec 05 10:40:43 crc kubenswrapper[4815]: I1205 10:40:43.237203 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-596686cb9d-rwjpw_3f638151-76e9-429e-b283-d741ad857276/barbican-keystone-listener-log/0.log" Dec 05 10:40:43 crc kubenswrapper[4815]: I1205 10:40:43.371557 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57db4d6df7-gdwss_92f1b7e3-1999-448d-90c1-9845b162701b/barbican-worker/0.log" Dec 05 10:40:43 crc kubenswrapper[4815]: I1205 10:40:43.551667 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-57db4d6df7-gdwss_92f1b7e3-1999-448d-90c1-9845b162701b/barbican-worker-log/0.log" Dec 05 10:40:43 crc kubenswrapper[4815]: I1205 10:40:43.667439 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-hdxl7_296de34a-3504-4fb8-8a20-02333f447f4c/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:43 crc kubenswrapper[4815]: I1205 10:40:43.810930 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_078b665f-5e54-4d83-966b-3684d06ea320/ceilometer-central-agent/0.log" Dec 05 10:40:43 crc kubenswrapper[4815]: I1205 10:40:43.868027 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_078b665f-5e54-4d83-966b-3684d06ea320/ceilometer-notification-agent/0.log" Dec 05 10:40:43 crc kubenswrapper[4815]: I1205 10:40:43.907799 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_078b665f-5e54-4d83-966b-3684d06ea320/proxy-httpd/0.log" Dec 05 10:40:43 crc kubenswrapper[4815]: I1205 10:40:43.968726 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_078b665f-5e54-4d83-966b-3684d06ea320/sg-core/0.log" Dec 05 10:40:44 crc kubenswrapper[4815]: I1205 10:40:44.095947 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-lm2r8_dd3bd6e3-0ff4-4ffb-a24c-532ef197e4f8/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:44 crc kubenswrapper[4815]: I1205 10:40:44.232036 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-5tl2r_18189644-5dd2-4329-a10b-70d62c5e0fd0/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:44 crc kubenswrapper[4815]: I1205 10:40:44.362611 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05/cinder-api/0.log" Dec 05 10:40:44 crc kubenswrapper[4815]: I1205 10:40:44.519907 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_b6d3ff6c-754a-4fbf-b363-c8ebe67f2e05/cinder-api-log/0.log" Dec 05 10:40:44 crc kubenswrapper[4815]: I1205 10:40:44.685249 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_6f53fc9d-0e41-4083-855a-1624fb520e84/probe/0.log" Dec 05 10:40:44 crc kubenswrapper[4815]: I1205 10:40:44.845236 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_6f53fc9d-0e41-4083-855a-1624fb520e84/cinder-backup/0.log" Dec 05 10:40:44 crc kubenswrapper[4815]: I1205 10:40:44.849698 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9a6d4017-ba13-47e9-b0e2-069a6d6afa21/cinder-scheduler/0.log" Dec 05 10:40:45 crc kubenswrapper[4815]: I1205 10:40:45.029778 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9a6d4017-ba13-47e9-b0e2-069a6d6afa21/probe/0.log" Dec 05 10:40:45 crc kubenswrapper[4815]: I1205 10:40:45.186919 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_15d9387d-aef1-48fe-8fb1-cb9d3dec22ea/probe/0.log" Dec 05 10:40:45 crc kubenswrapper[4815]: I1205 10:40:45.191325 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_15d9387d-aef1-48fe-8fb1-cb9d3dec22ea/cinder-volume/0.log" Dec 05 10:40:45 crc kubenswrapper[4815]: I1205 10:40:45.426021 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-6ddhh_4ab36e54-e39b-4b61-8609-52004d810c7a/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:45 crc kubenswrapper[4815]: I1205 10:40:45.471938 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-699sn_06fceca7-a988-4d36-8a56-fbe8fb96a28b/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:45 crc kubenswrapper[4815]: I1205 10:40:45.699147 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-94c764695-ml6b2_310bad94-91d6-4420-9d0c-8ed3af8b72c5/init/0.log" Dec 05 10:40:45 crc kubenswrapper[4815]: I1205 10:40:45.964274 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-94c764695-ml6b2_310bad94-91d6-4420-9d0c-8ed3af8b72c5/init/0.log" Dec 05 10:40:46 crc kubenswrapper[4815]: I1205 10:40:46.058162 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b/glance-httpd/0.log" Dec 05 10:40:46 crc kubenswrapper[4815]: I1205 10:40:46.130572 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-94c764695-ml6b2_310bad94-91d6-4420-9d0c-8ed3af8b72c5/dnsmasq-dns/0.log" Dec 05 10:40:46 crc kubenswrapper[4815]: I1205 10:40:46.213415 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_2dffe2c0-aaa7-47ff-bcd4-6cb1a055970b/glance-log/0.log" Dec 05 10:40:46 crc kubenswrapper[4815]: I1205 10:40:46.299137 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7330ea7b-3369-4ce7-95b6-fa415d3255f8/glance-httpd/0.log" Dec 05 10:40:46 crc kubenswrapper[4815]: I1205 10:40:46.369884 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7330ea7b-3369-4ce7-95b6-fa415d3255f8/glance-log/0.log" Dec 05 10:40:46 crc kubenswrapper[4815]: I1205 10:40:46.616026 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-74b8f449c4-9gwqr_e9604337-767f-492a-aa51-f31578f44ece/horizon/0.log" Dec 05 10:40:46 crc kubenswrapper[4815]: I1205 10:40:46.843908 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-jdgl9_abe1e909-bfe1-4bab-9b1e-5ed3316288b6/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:46 crc kubenswrapper[4815]: I1205 10:40:46.893118 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-74b8f449c4-9gwqr_e9604337-767f-492a-aa51-f31578f44ece/horizon-log/0.log" Dec 05 10:40:47 crc kubenswrapper[4815]: I1205 10:40:47.131520 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-cx6hg_711134c0-16f2-4ba4-8120-1c39b9d833b2/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:47 crc kubenswrapper[4815]: I1205 10:40:47.394236 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-6f56547bd4-2dxhd_7518e451-ec15-4df9-8fa0-842ca6094b36/keystone-api/0.log" Dec 05 10:40:47 crc kubenswrapper[4815]: I1205 10:40:47.407600 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29415481-7x47p_fba6bed7-9ef2-408d-ac41-04bff23ec468/keystone-cron/0.log" Dec 05 10:40:47 crc kubenswrapper[4815]: I1205 10:40:47.550510 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_b7f328c9-41c2-4d4d-9544-c22bcc026c33/kube-state-metrics/0.log" Dec 05 10:40:47 crc kubenswrapper[4815]: I1205 10:40:47.707609 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-kknjs_72f92018-4b16-474a-b5de-8fd124dd857b/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:47 crc kubenswrapper[4815]: I1205 10:40:47.915564 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_de4b5d94-2c11-4552-8230-08b776caa2a8/manila-api-log/0.log" Dec 05 10:40:47 crc kubenswrapper[4815]: I1205 10:40:47.942005 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_de4b5d94-2c11-4552-8230-08b776caa2a8/manila-api/0.log" Dec 05 10:40:48 crc kubenswrapper[4815]: I1205 10:40:48.112835 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_bed98f1a-2e43-47cc-82af-ca5c1a833995/probe/0.log" Dec 05 10:40:48 crc kubenswrapper[4815]: I1205 10:40:48.141305 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_bed98f1a-2e43-47cc-82af-ca5c1a833995/manila-scheduler/0.log" Dec 05 10:40:48 crc kubenswrapper[4815]: I1205 10:40:48.357796 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7d69d7b9-98dc-4dff-96bc-e562b8bc4b98/manila-share/0.log" Dec 05 10:40:48 crc kubenswrapper[4815]: I1205 10:40:48.375377 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_7d69d7b9-98dc-4dff-96bc-e562b8bc4b98/probe/0.log" Dec 05 10:40:48 crc kubenswrapper[4815]: I1205 10:40:48.757919 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54c4865f77-tppk8_fceefc10-15e5-4b68-9bb6-e786dde63941/neutron-httpd/0.log" Dec 05 10:40:48 crc kubenswrapper[4815]: I1205 10:40:48.907114 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-cpw6n_76992009-c2d1-4e3b-be35-70fc39e9f999/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:48 crc kubenswrapper[4815]: I1205 10:40:48.938897 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-54c4865f77-tppk8_fceefc10-15e5-4b68-9bb6-e786dde63941/neutron-api/0.log" Dec 05 10:40:49 crc kubenswrapper[4815]: I1205 10:40:49.637672 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_29c55b4a-b8d1-4604-ad9a-de77d453795f/nova-api-log/0.log" Dec 05 10:40:49 crc kubenswrapper[4815]: I1205 10:40:49.656961 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_7d358cc7-a297-4407-85db-ca619d6dccd4/nova-cell0-conductor-conductor/0.log" Dec 05 10:40:49 crc kubenswrapper[4815]: I1205 10:40:49.969345 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_62b5fcba-8ff7-4438-8b1c-960c0996a271/nova-cell1-conductor-conductor/0.log" Dec 05 10:40:50 crc kubenswrapper[4815]: I1205 10:40:50.117017 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_eea3bd17-9d01-4cf4-bdb2-4f5a48a25581/nova-cell1-novncproxy-novncproxy/0.log" Dec 05 10:40:50 crc kubenswrapper[4815]: I1205 10:40:50.165634 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_29c55b4a-b8d1-4604-ad9a-de77d453795f/nova-api-api/0.log" Dec 05 10:40:50 crc kubenswrapper[4815]: I1205 10:40:50.688729 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-xfxwl_e6418e66-6e27-4292-88c6-fca958ba3665/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:50 crc kubenswrapper[4815]: I1205 10:40:50.826855 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a7daef92-258a-4898-890a-44de7ad290b3/nova-metadata-log/0.log" Dec 05 10:40:51 crc kubenswrapper[4815]: I1205 10:40:51.289934 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1/mysql-bootstrap/0.log" Dec 05 10:40:51 crc kubenswrapper[4815]: I1205 10:40:51.319737 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_89bbf450-abe1-4daa-8060-40077543ac23/nova-scheduler-scheduler/0.log" Dec 05 10:40:51 crc kubenswrapper[4815]: I1205 10:40:51.576723 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1/galera/0.log" Dec 05 10:40:51 crc kubenswrapper[4815]: I1205 10:40:51.595141 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_8e67fbbb-c9b8-46c1-b761-c76a3f05dfe1/mysql-bootstrap/0.log" Dec 05 10:40:51 crc kubenswrapper[4815]: I1205 10:40:51.830032 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae4a421a-317a-4a4b-bcb1-2717aba4d25f/mysql-bootstrap/0.log" Dec 05 10:40:52 crc kubenswrapper[4815]: I1205 10:40:52.172082 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae4a421a-317a-4a4b-bcb1-2717aba4d25f/mysql-bootstrap/0.log" Dec 05 10:40:52 crc kubenswrapper[4815]: I1205 10:40:52.529575 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ae4a421a-317a-4a4b-bcb1-2717aba4d25f/galera/0.log" Dec 05 10:40:52 crc kubenswrapper[4815]: I1205 10:40:52.772401 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_04cdd6cd-6769-4296-a45d-fffd016fdfd6/openstackclient/0.log" Dec 05 10:40:52 crc kubenswrapper[4815]: I1205 10:40:52.783722 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-6q6fd_0ec613f7-a1c6-41d0-9d2d-5efb5284fdf0/openstack-network-exporter/0.log" Dec 05 10:40:52 crc kubenswrapper[4815]: I1205 10:40:52.981478 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a7daef92-258a-4898-890a-44de7ad290b3/nova-metadata-metadata/0.log" Dec 05 10:40:53 crc kubenswrapper[4815]: I1205 10:40:53.069063 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zlv59_6e811382-f977-41d3-aa18-65fe948261dd/ovsdb-server-init/0.log" Dec 05 10:40:53 crc kubenswrapper[4815]: I1205 10:40:53.359473 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zlv59_6e811382-f977-41d3-aa18-65fe948261dd/ovsdb-server-init/0.log" Dec 05 10:40:53 crc kubenswrapper[4815]: I1205 10:40:53.448860 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zlv59_6e811382-f977-41d3-aa18-65fe948261dd/ovs-vswitchd/0.log" Dec 05 10:40:53 crc kubenswrapper[4815]: I1205 10:40:53.449136 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zlv59_6e811382-f977-41d3-aa18-65fe948261dd/ovsdb-server/0.log" Dec 05 10:40:53 crc kubenswrapper[4815]: I1205 10:40:53.624479 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-v5mcb_0e0fc690-fb2a-4381-93df-1adcb21c0c7d/ovn-controller/0.log" Dec 05 10:40:53 crc kubenswrapper[4815]: I1205 10:40:53.767472 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-kv4qr_b3cdccbb-a413-4128-8d79-34ac79a10a45/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:54 crc kubenswrapper[4815]: I1205 10:40:54.047434 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c3e63e92-1982-4055-9df7-db07096a570e/openstack-network-exporter/0.log" Dec 05 10:40:54 crc kubenswrapper[4815]: I1205 10:40:54.203318 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c3e63e92-1982-4055-9df7-db07096a570e/ovn-northd/0.log" Dec 05 10:40:54 crc kubenswrapper[4815]: I1205 10:40:54.297232 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a065bb70-93f8-4e26-8ce0-6f92b7d7cba6/ovsdbserver-nb/0.log" Dec 05 10:40:54 crc kubenswrapper[4815]: I1205 10:40:54.362949 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a065bb70-93f8-4e26-8ce0-6f92b7d7cba6/openstack-network-exporter/0.log" Dec 05 10:40:54 crc kubenswrapper[4815]: I1205 10:40:54.645299 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_82494398-26f8-416f-88cd-65a7365cd2c2/openstack-network-exporter/0.log" Dec 05 10:40:54 crc kubenswrapper[4815]: I1205 10:40:54.783270 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_82494398-26f8-416f-88cd-65a7365cd2c2/ovsdbserver-sb/0.log" Dec 05 10:40:55 crc kubenswrapper[4815]: I1205 10:40:55.090149 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-549777fc4b-mjkbk_75817271-e950-4196-bc2c-b32ac4dad684/placement-api/0.log" Dec 05 10:40:55 crc kubenswrapper[4815]: I1205 10:40:55.112871 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_db8bc583-6bcf-40c7-875b-d5c9544af543/setup-container/0.log" Dec 05 10:40:55 crc kubenswrapper[4815]: I1205 10:40:55.296601 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-549777fc4b-mjkbk_75817271-e950-4196-bc2c-b32ac4dad684/placement-log/0.log" Dec 05 10:40:55 crc kubenswrapper[4815]: I1205 10:40:55.681415 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_db8bc583-6bcf-40c7-875b-d5c9544af543/rabbitmq/0.log" Dec 05 10:40:55 crc kubenswrapper[4815]: I1205 10:40:55.714890 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d0d7070e-5d74-4d40-9fe9-b42c60585002/setup-container/0.log" Dec 05 10:40:55 crc kubenswrapper[4815]: I1205 10:40:55.982339 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_db8bc583-6bcf-40c7-875b-d5c9544af543/setup-container/0.log" Dec 05 10:40:56 crc kubenswrapper[4815]: I1205 10:40:56.339075 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d0d7070e-5d74-4d40-9fe9-b42c60585002/setup-container/0.log" Dec 05 10:40:56 crc kubenswrapper[4815]: I1205 10:40:56.452143 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d0d7070e-5d74-4d40-9fe9-b42c60585002/rabbitmq/0.log" Dec 05 10:40:56 crc kubenswrapper[4815]: I1205 10:40:56.580351 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-nghzz_e775d3b4-8e2f-4255-8175-9664129b7998/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:56 crc kubenswrapper[4815]: I1205 10:40:56.706977 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-lsj5s_0e0848e7-a6e0-4672-bb56-2f40193be881/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:57 crc kubenswrapper[4815]: I1205 10:40:57.061359 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-46rkt_7251f2d6-ee27-4b5d-9d1a-6aeab813638b/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:40:57 crc kubenswrapper[4815]: I1205 10:40:57.127770 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-w8b66_ba6c4e35-e275-4cd3-8f30-0102966dcc1b/ssh-known-hosts-edpm-deployment/0.log" Dec 05 10:40:57 crc kubenswrapper[4815]: I1205 10:40:57.356288 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_b0ec9924-8223-4a07-806b-0ee6451a7021/tempest-tests-tempest-tests-runner/0.log" Dec 05 10:40:57 crc kubenswrapper[4815]: I1205 10:40:57.478249 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_6fb2333c-6791-4164-a041-fec86a88cc06/test-operator-logs-container/0.log" Dec 05 10:40:57 crc kubenswrapper[4815]: I1205 10:40:57.672766 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-f54t2_abc71b71-d4f3-4c99-8a40-8969b4371f39/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 05 10:41:12 crc kubenswrapper[4815]: I1205 10:41:12.737468 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_28f58061-0d04-4513-96f0-7d221940a154/memcached/0.log" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.509391 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xkl26"] Dec 05 10:41:30 crc kubenswrapper[4815]: E1205 10:41:30.510530 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56156298-bd80-41f4-bbe7-6c0603005afd" containerName="registry-server" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.510561 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="56156298-bd80-41f4-bbe7-6c0603005afd" containerName="registry-server" Dec 05 10:41:30 crc kubenswrapper[4815]: E1205 10:41:30.510598 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56156298-bd80-41f4-bbe7-6c0603005afd" containerName="extract-content" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.510607 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="56156298-bd80-41f4-bbe7-6c0603005afd" containerName="extract-content" Dec 05 10:41:30 crc kubenswrapper[4815]: E1205 10:41:30.510634 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56156298-bd80-41f4-bbe7-6c0603005afd" containerName="extract-utilities" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.510642 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="56156298-bd80-41f4-bbe7-6c0603005afd" containerName="extract-utilities" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.510903 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="56156298-bd80-41f4-bbe7-6c0603005afd" containerName="registry-server" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.512462 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.549625 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xkl26"] Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.627859 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-catalog-content\") pod \"community-operators-xkl26\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.627919 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d5cl\" (UniqueName: \"kubernetes.io/projected/922e33c0-945e-42eb-a0c6-b7a0262e9563-kube-api-access-7d5cl\") pod \"community-operators-xkl26\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.628271 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-utilities\") pod \"community-operators-xkl26\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.730897 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-catalog-content\") pod \"community-operators-xkl26\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.730965 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7d5cl\" (UniqueName: \"kubernetes.io/projected/922e33c0-945e-42eb-a0c6-b7a0262e9563-kube-api-access-7d5cl\") pod \"community-operators-xkl26\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.731063 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-utilities\") pod \"community-operators-xkl26\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.731578 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-catalog-content\") pod \"community-operators-xkl26\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.731694 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-utilities\") pod \"community-operators-xkl26\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.774821 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d5cl\" (UniqueName: \"kubernetes.io/projected/922e33c0-945e-42eb-a0c6-b7a0262e9563-kube-api-access-7d5cl\") pod \"community-operators-xkl26\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:30 crc kubenswrapper[4815]: I1205 10:41:30.850769 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:31 crc kubenswrapper[4815]: I1205 10:41:31.433570 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xkl26"] Dec 05 10:41:31 crc kubenswrapper[4815]: I1205 10:41:31.443774 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xkl26" event={"ID":"922e33c0-945e-42eb-a0c6-b7a0262e9563","Type":"ContainerStarted","Data":"7ea968fb0d748e6720df183def3eef0eef5fa35aae2f8efb8ae62f03ca70f08d"} Dec 05 10:41:32 crc kubenswrapper[4815]: I1205 10:41:32.454022 4815 generic.go:334] "Generic (PLEG): container finished" podID="922e33c0-945e-42eb-a0c6-b7a0262e9563" containerID="56d56787adc2ebea43f66225b9c62e3c67be33cb4e197e5f409c71556982c33b" exitCode=0 Dec 05 10:41:32 crc kubenswrapper[4815]: I1205 10:41:32.454105 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xkl26" event={"ID":"922e33c0-945e-42eb-a0c6-b7a0262e9563","Type":"ContainerDied","Data":"56d56787adc2ebea43f66225b9c62e3c67be33cb4e197e5f409c71556982c33b"} Dec 05 10:41:33 crc kubenswrapper[4815]: I1205 10:41:33.463768 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xkl26" event={"ID":"922e33c0-945e-42eb-a0c6-b7a0262e9563","Type":"ContainerStarted","Data":"b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213"} Dec 05 10:41:33 crc kubenswrapper[4815]: I1205 10:41:33.539731 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/util/0.log" Dec 05 10:41:33 crc kubenswrapper[4815]: I1205 10:41:33.769370 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/util/0.log" Dec 05 10:41:33 crc kubenswrapper[4815]: I1205 10:41:33.857517 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/pull/0.log" Dec 05 10:41:33 crc kubenswrapper[4815]: I1205 10:41:33.969416 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/pull/0.log" Dec 05 10:41:34 crc kubenswrapper[4815]: I1205 10:41:34.406033 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/pull/0.log" Dec 05 10:41:34 crc kubenswrapper[4815]: I1205 10:41:34.551645 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/extract/0.log" Dec 05 10:41:34 crc kubenswrapper[4815]: I1205 10:41:34.600028 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_917aae072417a6c2fc5ddd97ca05bfedb9fc1cad89a3b1c4d989b78eafx7qfz_e115064e-50f8-4b07-962d-fa7e7ea7108f/util/0.log" Dec 05 10:41:34 crc kubenswrapper[4815]: I1205 10:41:34.659807 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-26tt9_cafb574b-a659-41e6-89d5-c02851274f73/kube-rbac-proxy/0.log" Dec 05 10:41:34 crc kubenswrapper[4815]: I1205 10:41:34.781183 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-j8cxc_f023493d-1716-40ac-816d-26067f0019f3/kube-rbac-proxy/0.log" Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.020318 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-26tt9_cafb574b-a659-41e6-89d5-c02851274f73/manager/0.log" Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.023515 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-j8cxc_f023493d-1716-40ac-816d-26067f0019f3/manager/0.log" Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.096453 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-g4nbp_9db099f9-ed44-46ed-ab3d-97d09ee0b5b0/kube-rbac-proxy/0.log" Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.266682 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-g4nbp_9db099f9-ed44-46ed-ab3d-97d09ee0b5b0/manager/0.log" Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.433724 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-8blh4_591bafca-5706-418f-b803-d29d1b0865db/kube-rbac-proxy/0.log" Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.446920 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-8blh4_591bafca-5706-418f-b803-d29d1b0865db/manager/0.log" Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.482118 4815 generic.go:334] "Generic (PLEG): container finished" podID="922e33c0-945e-42eb-a0c6-b7a0262e9563" containerID="b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213" exitCode=0 Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.482167 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xkl26" event={"ID":"922e33c0-945e-42eb-a0c6-b7a0262e9563","Type":"ContainerDied","Data":"b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213"} Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.625237 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-tmg9c_f23f3b0d-7874-4906-81de-b8d9226082d2/kube-rbac-proxy/0.log" Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.650440 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-tmg9c_f23f3b0d-7874-4906-81de-b8d9226082d2/manager/0.log" Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.849458 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5jh5q_ed463fa2-6734-41b5-a34b-cb83436130fd/kube-rbac-proxy/0.log" Dec 05 10:41:35 crc kubenswrapper[4815]: I1205 10:41:35.884926 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-5jh5q_ed463fa2-6734-41b5-a34b-cb83436130fd/manager/0.log" Dec 05 10:41:36 crc kubenswrapper[4815]: I1205 10:41:36.105423 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-9pflh_4abce90f-180f-4302-aecd-89bf5739757c/kube-rbac-proxy/0.log" Dec 05 10:41:36 crc kubenswrapper[4815]: I1205 10:41:36.297813 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-9pflh_4abce90f-180f-4302-aecd-89bf5739757c/manager/0.log" Dec 05 10:41:36 crc kubenswrapper[4815]: I1205 10:41:36.430756 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7mtff_ef1e93b5-9a31-4588-a5e4-78dbe6ef0437/kube-rbac-proxy/0.log" Dec 05 10:41:36 crc kubenswrapper[4815]: I1205 10:41:36.602063 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xkl26" event={"ID":"922e33c0-945e-42eb-a0c6-b7a0262e9563","Type":"ContainerStarted","Data":"870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b"} Dec 05 10:41:36 crc kubenswrapper[4815]: I1205 10:41:36.627797 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xkl26" podStartSLOduration=3.213133147 podStartE2EDuration="6.627770793s" podCreationTimestamp="2025-12-05 10:41:30 +0000 UTC" firstStartedPulling="2025-12-05 10:41:32.457097071 +0000 UTC m=+5691.335703908" lastFinishedPulling="2025-12-05 10:41:35.871734717 +0000 UTC m=+5694.750341554" observedRunningTime="2025-12-05 10:41:36.625734788 +0000 UTC m=+5695.504341625" watchObservedRunningTime="2025-12-05 10:41:36.627770793 +0000 UTC m=+5695.506377630" Dec 05 10:41:36 crc kubenswrapper[4815]: I1205 10:41:36.719992 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-7mtff_ef1e93b5-9a31-4588-a5e4-78dbe6ef0437/manager/0.log" Dec 05 10:41:36 crc kubenswrapper[4815]: I1205 10:41:36.871301 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-w7vjm_cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb/kube-rbac-proxy/0.log" Dec 05 10:41:37 crc kubenswrapper[4815]: I1205 10:41:37.417889 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-w7vjm_cd4a58e6-01e1-4c5d-94a6-f20439f2f7eb/manager/0.log" Dec 05 10:41:37 crc kubenswrapper[4815]: I1205 10:41:37.701098 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-2k8wt_f80008bd-769f-4960-be8f-62894fdd7718/kube-rbac-proxy/0.log" Dec 05 10:41:37 crc kubenswrapper[4815]: I1205 10:41:37.703366 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-2k8wt_f80008bd-769f-4960-be8f-62894fdd7718/manager/0.log" Dec 05 10:41:37 crc kubenswrapper[4815]: I1205 10:41:37.849845 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-bm8kq_3665adde-25ad-4261-be50-ef4e21726b7b/kube-rbac-proxy/0.log" Dec 05 10:41:38 crc kubenswrapper[4815]: I1205 10:41:38.004694 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-bm8kq_3665adde-25ad-4261-be50-ef4e21726b7b/manager/0.log" Dec 05 10:41:38 crc kubenswrapper[4815]: I1205 10:41:38.194652 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-ccd95_7482e78f-1387-4e91-b5d4-419f5164fea5/kube-rbac-proxy/0.log" Dec 05 10:41:38 crc kubenswrapper[4815]: I1205 10:41:38.224438 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-ccd95_7482e78f-1387-4e91-b5d4-419f5164fea5/manager/0.log" Dec 05 10:41:38 crc kubenswrapper[4815]: I1205 10:41:38.380616 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-25lnj_48f9720a-5d73-4773-85f9-9c571c9322ee/kube-rbac-proxy/0.log" Dec 05 10:41:38 crc kubenswrapper[4815]: I1205 10:41:38.548465 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-25lnj_48f9720a-5d73-4773-85f9-9c571c9322ee/manager/0.log" Dec 05 10:41:38 crc kubenswrapper[4815]: I1205 10:41:38.787611 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-l5qqd_b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d/kube-rbac-proxy/0.log" Dec 05 10:41:38 crc kubenswrapper[4815]: I1205 10:41:38.877651 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-l5qqd_b25b6d76-6c06-45c1-9510-eb3f3aeb8a6d/manager/0.log" Dec 05 10:41:39 crc kubenswrapper[4815]: I1205 10:41:39.080501 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55c85496f5b4xkh_b7558b70-107c-48ea-ac22-e42b1bcdf47d/manager/0.log" Dec 05 10:41:39 crc kubenswrapper[4815]: I1205 10:41:39.107017 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-55c85496f5b4xkh_b7558b70-107c-48ea-ac22-e42b1bcdf47d/kube-rbac-proxy/0.log" Dec 05 10:41:39 crc kubenswrapper[4815]: I1205 10:41:39.584944 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-55b6fb9447-xg6vd_7f520180-b5ab-420b-9479-4089e1845ecc/operator/0.log" Dec 05 10:41:39 crc kubenswrapper[4815]: I1205 10:41:39.695365 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xttqw_6943ef97-36ce-4f74-b31b-4ba90192d3ac/registry-server/0.log" Dec 05 10:41:39 crc kubenswrapper[4815]: I1205 10:41:39.751654 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-5jlqr_378923e7-18f6-4946-8e86-d1ececca15cc/kube-rbac-proxy/0.log" Dec 05 10:41:39 crc kubenswrapper[4815]: I1205 10:41:39.858801 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-5jlqr_378923e7-18f6-4946-8e86-d1ececca15cc/manager/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.027070 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-8pt5c_6b41117c-2922-4337-b00c-2bebd57a2c6b/manager/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.108143 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-8pt5c_6b41117c-2922-4337-b00c-2bebd57a2c6b/kube-rbac-proxy/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.235552 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-mvkws_888d282d-8aa1-4a36-9286-d00af63950ba/operator/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.345843 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-54bdf956c4-2lknm_8f0cbf2b-67d6-4058-b173-d24117dfcdd7/manager/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.405246 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6nl6f_47664970-bfba-453a-ad06-ac1e03417eac/kube-rbac-proxy/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.450289 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-6nl6f_47664970-bfba-453a-ad06-ac1e03417eac/manager/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.640968 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bnb4c_926f3f91-cee1-42ae-888f-73ec7f20dd61/kube-rbac-proxy/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.666333 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-76cc84c6bb-bnb4c_926f3f91-cee1-42ae-888f-73ec7f20dd61/manager/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.781160 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-gbkbf_d8b41c8b-030e-4082-98a3-a0e9fc05a208/kube-rbac-proxy/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.830095 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-gbkbf_d8b41c8b-030e-4082-98a3-a0e9fc05a208/manager/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.851728 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.851795 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.910452 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.984143 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-j4r7f_19f20867-5315-4be3-836b-57bb0b501b36/kube-rbac-proxy/0.log" Dec 05 10:41:40 crc kubenswrapper[4815]: I1205 10:41:40.992400 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-j4r7f_19f20867-5315-4be3-836b-57bb0b501b36/manager/0.log" Dec 05 10:41:41 crc kubenswrapper[4815]: I1205 10:41:41.763430 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:41 crc kubenswrapper[4815]: I1205 10:41:41.824351 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xkl26"] Dec 05 10:41:43 crc kubenswrapper[4815]: I1205 10:41:43.659590 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xkl26" podUID="922e33c0-945e-42eb-a0c6-b7a0262e9563" containerName="registry-server" containerID="cri-o://870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b" gracePeriod=2 Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.311728 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.509704 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-catalog-content\") pod \"922e33c0-945e-42eb-a0c6-b7a0262e9563\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.509935 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-utilities\") pod \"922e33c0-945e-42eb-a0c6-b7a0262e9563\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.510191 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7d5cl\" (UniqueName: \"kubernetes.io/projected/922e33c0-945e-42eb-a0c6-b7a0262e9563-kube-api-access-7d5cl\") pod \"922e33c0-945e-42eb-a0c6-b7a0262e9563\" (UID: \"922e33c0-945e-42eb-a0c6-b7a0262e9563\") " Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.513144 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-utilities" (OuterVolumeSpecName: "utilities") pod "922e33c0-945e-42eb-a0c6-b7a0262e9563" (UID: "922e33c0-945e-42eb-a0c6-b7a0262e9563"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.514119 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.541043 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/922e33c0-945e-42eb-a0c6-b7a0262e9563-kube-api-access-7d5cl" (OuterVolumeSpecName: "kube-api-access-7d5cl") pod "922e33c0-945e-42eb-a0c6-b7a0262e9563" (UID: "922e33c0-945e-42eb-a0c6-b7a0262e9563"). InnerVolumeSpecName "kube-api-access-7d5cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.575331 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "922e33c0-945e-42eb-a0c6-b7a0262e9563" (UID: "922e33c0-945e-42eb-a0c6-b7a0262e9563"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.616373 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7d5cl\" (UniqueName: \"kubernetes.io/projected/922e33c0-945e-42eb-a0c6-b7a0262e9563-kube-api-access-7d5cl\") on node \"crc\" DevicePath \"\"" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.616410 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/922e33c0-945e-42eb-a0c6-b7a0262e9563-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.670682 4815 generic.go:334] "Generic (PLEG): container finished" podID="922e33c0-945e-42eb-a0c6-b7a0262e9563" containerID="870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b" exitCode=0 Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.670728 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xkl26" event={"ID":"922e33c0-945e-42eb-a0c6-b7a0262e9563","Type":"ContainerDied","Data":"870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b"} Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.670760 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xkl26" event={"ID":"922e33c0-945e-42eb-a0c6-b7a0262e9563","Type":"ContainerDied","Data":"7ea968fb0d748e6720df183def3eef0eef5fa35aae2f8efb8ae62f03ca70f08d"} Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.670791 4815 scope.go:117] "RemoveContainer" containerID="870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.670812 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xkl26" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.691263 4815 scope.go:117] "RemoveContainer" containerID="b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.718117 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xkl26"] Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.725037 4815 scope.go:117] "RemoveContainer" containerID="56d56787adc2ebea43f66225b9c62e3c67be33cb4e197e5f409c71556982c33b" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.726720 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xkl26"] Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.776176 4815 scope.go:117] "RemoveContainer" containerID="870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b" Dec 05 10:41:44 crc kubenswrapper[4815]: E1205 10:41:44.778447 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b\": container with ID starting with 870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b not found: ID does not exist" containerID="870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.778484 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b"} err="failed to get container status \"870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b\": rpc error: code = NotFound desc = could not find container \"870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b\": container with ID starting with 870a234baae60fe2765774665badfda0738a8795d1581e1b5cf001fd6daa1a1b not found: ID does not exist" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.778522 4815 scope.go:117] "RemoveContainer" containerID="b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213" Dec 05 10:41:44 crc kubenswrapper[4815]: E1205 10:41:44.779232 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213\": container with ID starting with b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213 not found: ID does not exist" containerID="b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.779279 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213"} err="failed to get container status \"b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213\": rpc error: code = NotFound desc = could not find container \"b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213\": container with ID starting with b3bd57e19bc6b4288bac87a9ee5dbb4d256d1988276a680a8fccf7508870f213 not found: ID does not exist" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.779308 4815 scope.go:117] "RemoveContainer" containerID="56d56787adc2ebea43f66225b9c62e3c67be33cb4e197e5f409c71556982c33b" Dec 05 10:41:44 crc kubenswrapper[4815]: E1205 10:41:44.779662 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56d56787adc2ebea43f66225b9c62e3c67be33cb4e197e5f409c71556982c33b\": container with ID starting with 56d56787adc2ebea43f66225b9c62e3c67be33cb4e197e5f409c71556982c33b not found: ID does not exist" containerID="56d56787adc2ebea43f66225b9c62e3c67be33cb4e197e5f409c71556982c33b" Dec 05 10:41:44 crc kubenswrapper[4815]: I1205 10:41:44.779692 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56d56787adc2ebea43f66225b9c62e3c67be33cb4e197e5f409c71556982c33b"} err="failed to get container status \"56d56787adc2ebea43f66225b9c62e3c67be33cb4e197e5f409c71556982c33b\": rpc error: code = NotFound desc = could not find container \"56d56787adc2ebea43f66225b9c62e3c67be33cb4e197e5f409c71556982c33b\": container with ID starting with 56d56787adc2ebea43f66225b9c62e3c67be33cb4e197e5f409c71556982c33b not found: ID does not exist" Dec 05 10:41:45 crc kubenswrapper[4815]: I1205 10:41:45.431798 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="922e33c0-945e-42eb-a0c6-b7a0262e9563" path="/var/lib/kubelet/pods/922e33c0-945e-42eb-a0c6-b7a0262e9563/volumes" Dec 05 10:42:04 crc kubenswrapper[4815]: I1205 10:42:04.423114 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-h89pl_fb48516f-711a-4131-a555-de5ee6e96ae1/control-plane-machine-set-operator/0.log" Dec 05 10:42:04 crc kubenswrapper[4815]: I1205 10:42:04.591999 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tw877_afd36151-d969-4642-b6c4-c6b3b5f65476/kube-rbac-proxy/0.log" Dec 05 10:42:04 crc kubenswrapper[4815]: I1205 10:42:04.698050 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-tw877_afd36151-d969-4642-b6c4-c6b3b5f65476/machine-api-operator/0.log" Dec 05 10:42:18 crc kubenswrapper[4815]: I1205 10:42:18.597346 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-vcstd_82c69bfe-d893-41ec-bc5d-d872d51c8c15/cert-manager-controller/0.log" Dec 05 10:42:18 crc kubenswrapper[4815]: I1205 10:42:18.782331 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-zc6mp_8f3da4f1-671e-4911-a481-45824ce433d5/cert-manager-cainjector/0.log" Dec 05 10:42:18 crc kubenswrapper[4815]: I1205 10:42:18.890440 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-sz66k_56e3935a-c7dd-4bb2-9dff-deaf3dd7e25b/cert-manager-webhook/0.log" Dec 05 10:42:33 crc kubenswrapper[4815]: I1205 10:42:33.021295 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-5g45b_028ef3e8-a0e1-443e-840d-0f7244a0e953/nmstate-console-plugin/0.log" Dec 05 10:42:33 crc kubenswrapper[4815]: I1205 10:42:33.273445 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-d8hkx_da8c41e3-2e9f-4917-a9e7-efa4debc22c7/nmstate-handler/0.log" Dec 05 10:42:33 crc kubenswrapper[4815]: I1205 10:42:33.290169 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9frbq_d47f3938-b642-4bd2-b807-b2c962918651/kube-rbac-proxy/0.log" Dec 05 10:42:33 crc kubenswrapper[4815]: I1205 10:42:33.333077 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-9frbq_d47f3938-b642-4bd2-b807-b2c962918651/nmstate-metrics/0.log" Dec 05 10:42:33 crc kubenswrapper[4815]: I1205 10:42:33.506753 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-82r92_7e7d72c9-18b4-4716-aabb-2c55c7070020/nmstate-operator/0.log" Dec 05 10:42:33 crc kubenswrapper[4815]: I1205 10:42:33.625296 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-qtprq_8cfb92c5-84f6-4b2d-a736-69c10b8ffdb5/nmstate-webhook/0.log" Dec 05 10:42:49 crc kubenswrapper[4815]: I1205 10:42:49.899516 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-j8cbz_55c34ec0-bd62-4eb5-a838-48e0622ebea9/kube-rbac-proxy/0.log" Dec 05 10:42:49 crc kubenswrapper[4815]: I1205 10:42:49.937960 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-j8cbz_55c34ec0-bd62-4eb5-a838-48e0622ebea9/controller/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.121304 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-frr-files/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.192180 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.192243 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.328483 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-metrics/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.329876 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-frr-files/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.368971 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-reloader/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.380115 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-reloader/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.610157 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-metrics/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.658763 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-metrics/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.661045 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-frr-files/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.670935 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-reloader/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.857306 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-frr-files/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.864608 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/controller/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.896309 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-metrics/0.log" Dec 05 10:42:50 crc kubenswrapper[4815]: I1205 10:42:50.899143 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/cp-reloader/0.log" Dec 05 10:42:51 crc kubenswrapper[4815]: I1205 10:42:51.053027 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/frr-metrics/0.log" Dec 05 10:42:51 crc kubenswrapper[4815]: I1205 10:42:51.108084 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/kube-rbac-proxy/0.log" Dec 05 10:42:51 crc kubenswrapper[4815]: I1205 10:42:51.146811 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/kube-rbac-proxy-frr/0.log" Dec 05 10:42:51 crc kubenswrapper[4815]: I1205 10:42:51.600072 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/reloader/0.log" Dec 05 10:42:51 crc kubenswrapper[4815]: I1205 10:42:51.740219 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-b6cnn_46274b62-2000-4c62-8688-af6b5fcb87dc/frr-k8s-webhook-server/0.log" Dec 05 10:42:51 crc kubenswrapper[4815]: I1205 10:42:51.928369 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-7f5859557b-zxp74_37e3bb94-27f2-4bfe-9c26-4bbb3025330c/manager/0.log" Dec 05 10:42:52 crc kubenswrapper[4815]: I1205 10:42:52.092257 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-69858568cc-scmsp_1aabbb7d-8736-4ff3-a41e-7599a4e74194/webhook-server/0.log" Dec 05 10:42:52 crc kubenswrapper[4815]: I1205 10:42:52.386835 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xszxj_a72253a0-2394-424d-8689-1c989ce6b2ed/kube-rbac-proxy/0.log" Dec 05 10:42:52 crc kubenswrapper[4815]: I1205 10:42:52.876451 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-xszxj_a72253a0-2394-424d-8689-1c989ce6b2ed/speaker/0.log" Dec 05 10:42:52 crc kubenswrapper[4815]: I1205 10:42:52.946136 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jbdfv_563ef3a0-d97f-4b5b-afc4-378bdbbee19b/frr/0.log" Dec 05 10:43:05 crc kubenswrapper[4815]: I1205 10:43:05.680123 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/util/0.log" Dec 05 10:43:05 crc kubenswrapper[4815]: I1205 10:43:05.919891 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/pull/0.log" Dec 05 10:43:05 crc kubenswrapper[4815]: I1205 10:43:05.942103 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/pull/0.log" Dec 05 10:43:05 crc kubenswrapper[4815]: I1205 10:43:05.971639 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/util/0.log" Dec 05 10:43:06 crc kubenswrapper[4815]: I1205 10:43:06.153915 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/pull/0.log" Dec 05 10:43:06 crc kubenswrapper[4815]: I1205 10:43:06.164675 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/extract/0.log" Dec 05 10:43:06 crc kubenswrapper[4815]: I1205 10:43:06.198376 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fqzffp_f6f8e345-b259-4c23-a98a-22ff6454d46d/util/0.log" Dec 05 10:43:06 crc kubenswrapper[4815]: I1205 10:43:06.344347 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/util/0.log" Dec 05 10:43:06 crc kubenswrapper[4815]: I1205 10:43:06.519977 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/util/0.log" Dec 05 10:43:06 crc kubenswrapper[4815]: I1205 10:43:06.529793 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/pull/0.log" Dec 05 10:43:06 crc kubenswrapper[4815]: I1205 10:43:06.550220 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/pull/0.log" Dec 05 10:43:06 crc kubenswrapper[4815]: I1205 10:43:06.774887 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/pull/0.log" Dec 05 10:43:06 crc kubenswrapper[4815]: I1205 10:43:06.807387 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/extract/0.log" Dec 05 10:43:06 crc kubenswrapper[4815]: I1205 10:43:06.864335 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83ccbzw_1f7f5e67-9cbf-4059-9afd-ea5224d6e571/util/0.log" Dec 05 10:43:07 crc kubenswrapper[4815]: I1205 10:43:07.091126 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-utilities/0.log" Dec 05 10:43:07 crc kubenswrapper[4815]: I1205 10:43:07.297180 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-content/0.log" Dec 05 10:43:07 crc kubenswrapper[4815]: I1205 10:43:07.299406 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-content/0.log" Dec 05 10:43:07 crc kubenswrapper[4815]: I1205 10:43:07.299881 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-utilities/0.log" Dec 05 10:43:07 crc kubenswrapper[4815]: I1205 10:43:07.547137 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-content/0.log" Dec 05 10:43:07 crc kubenswrapper[4815]: I1205 10:43:07.557868 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/extract-utilities/0.log" Dec 05 10:43:07 crc kubenswrapper[4815]: I1205 10:43:07.953477 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-utilities/0.log" Dec 05 10:43:08 crc kubenswrapper[4815]: I1205 10:43:08.231019 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-utilities/0.log" Dec 05 10:43:08 crc kubenswrapper[4815]: I1205 10:43:08.258085 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-content/0.log" Dec 05 10:43:08 crc kubenswrapper[4815]: I1205 10:43:08.419398 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-content/0.log" Dec 05 10:43:08 crc kubenswrapper[4815]: I1205 10:43:08.430477 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-l4v4h_50a2f4af-33e9-4508-853b-6142ca44e9ef/registry-server/0.log" Dec 05 10:43:08 crc kubenswrapper[4815]: I1205 10:43:08.577336 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-content/0.log" Dec 05 10:43:08 crc kubenswrapper[4815]: I1205 10:43:08.586752 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/extract-utilities/0.log" Dec 05 10:43:09 crc kubenswrapper[4815]: I1205 10:43:09.055616 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-6k7mn_a4c4b634-bebd-41c4-9462-d33c8a9ff1cd/marketplace-operator/0.log" Dec 05 10:43:09 crc kubenswrapper[4815]: I1205 10:43:09.067790 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-utilities/0.log" Dec 05 10:43:09 crc kubenswrapper[4815]: I1205 10:43:09.307008 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-utilities/0.log" Dec 05 10:43:09 crc kubenswrapper[4815]: I1205 10:43:09.405304 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-content/0.log" Dec 05 10:43:09 crc kubenswrapper[4815]: I1205 10:43:09.441624 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-content/0.log" Dec 05 10:43:09 crc kubenswrapper[4815]: I1205 10:43:09.522229 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-ssfkz_f6be68d8-ee10-4dad-b1d5-5f1d7826917b/registry-server/0.log" Dec 05 10:43:09 crc kubenswrapper[4815]: I1205 10:43:09.649595 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-utilities/0.log" Dec 05 10:43:09 crc kubenswrapper[4815]: I1205 10:43:09.697766 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/extract-content/0.log" Dec 05 10:43:09 crc kubenswrapper[4815]: I1205 10:43:09.900049 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-4lpmz_d66535f7-8f1a-48cd-af4d-5c6a41535fe5/registry-server/0.log" Dec 05 10:43:09 crc kubenswrapper[4815]: I1205 10:43:09.921182 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-utilities/0.log" Dec 05 10:43:10 crc kubenswrapper[4815]: I1205 10:43:10.200719 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-content/0.log" Dec 05 10:43:10 crc kubenswrapper[4815]: I1205 10:43:10.237130 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-utilities/0.log" Dec 05 10:43:10 crc kubenswrapper[4815]: I1205 10:43:10.341779 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-content/0.log" Dec 05 10:43:10 crc kubenswrapper[4815]: I1205 10:43:10.452658 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-content/0.log" Dec 05 10:43:10 crc kubenswrapper[4815]: I1205 10:43:10.491661 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/extract-utilities/0.log" Dec 05 10:43:11 crc kubenswrapper[4815]: I1205 10:43:11.056797 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-t2dbx_18bd7707-24bd-4562-a633-c24cd47e50d0/registry-server/0.log" Dec 05 10:43:20 crc kubenswrapper[4815]: I1205 10:43:20.192178 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:43:20 crc kubenswrapper[4815]: I1205 10:43:20.192772 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:43:50 crc kubenswrapper[4815]: I1205 10:43:50.192739 4815 patch_prober.go:28] interesting pod/machine-config-daemon-4wq7d container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 05 10:43:50 crc kubenswrapper[4815]: I1205 10:43:50.194272 4815 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 05 10:43:50 crc kubenswrapper[4815]: I1205 10:43:50.194404 4815 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" Dec 05 10:43:50 crc kubenswrapper[4815]: I1205 10:43:50.195231 4815 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94"} pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 05 10:43:50 crc kubenswrapper[4815]: I1205 10:43:50.195418 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" containerName="machine-config-daemon" containerID="cri-o://4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" gracePeriod=600 Dec 05 10:43:50 crc kubenswrapper[4815]: E1205 10:43:50.406685 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:43:50 crc kubenswrapper[4815]: E1205 10:43:50.491291 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0a874c44_831d_4f81_bf48_2025f5615574.slice/crio-conmon-4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94.scope\": RecentStats: unable to find data in memory cache]" Dec 05 10:43:50 crc kubenswrapper[4815]: I1205 10:43:50.824117 4815 generic.go:334] "Generic (PLEG): container finished" podID="0a874c44-831d-4f81-bf48-2025f5615574" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" exitCode=0 Dec 05 10:43:50 crc kubenswrapper[4815]: I1205 10:43:50.824288 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerDied","Data":"4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94"} Dec 05 10:43:50 crc kubenswrapper[4815]: I1205 10:43:50.824442 4815 scope.go:117] "RemoveContainer" containerID="36532cd36fe5a7fe6381b37918506e1e1d6db7ef1721fc2254713e8eb9050c93" Dec 05 10:43:50 crc kubenswrapper[4815]: I1205 10:43:50.825090 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:43:50 crc kubenswrapper[4815]: E1205 10:43:50.825360 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:44:03 crc kubenswrapper[4815]: I1205 10:44:03.418774 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:44:03 crc kubenswrapper[4815]: E1205 10:44:03.419752 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:44:17 crc kubenswrapper[4815]: I1205 10:44:17.419735 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:44:17 crc kubenswrapper[4815]: E1205 10:44:17.420408 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:44:30 crc kubenswrapper[4815]: I1205 10:44:30.421194 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:44:30 crc kubenswrapper[4815]: E1205 10:44:30.422723 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:44:43 crc kubenswrapper[4815]: I1205 10:44:43.418912 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:44:43 crc kubenswrapper[4815]: E1205 10:44:43.420225 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:44:55 crc kubenswrapper[4815]: I1205 10:44:55.419044 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:44:55 crc kubenswrapper[4815]: E1205 10:44:55.419702 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.220168 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn"] Dec 05 10:45:00 crc kubenswrapper[4815]: E1205 10:45:00.221324 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="922e33c0-945e-42eb-a0c6-b7a0262e9563" containerName="registry-server" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.221353 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="922e33c0-945e-42eb-a0c6-b7a0262e9563" containerName="registry-server" Dec 05 10:45:00 crc kubenswrapper[4815]: E1205 10:45:00.221372 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="922e33c0-945e-42eb-a0c6-b7a0262e9563" containerName="extract-content" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.221381 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="922e33c0-945e-42eb-a0c6-b7a0262e9563" containerName="extract-content" Dec 05 10:45:00 crc kubenswrapper[4815]: E1205 10:45:00.221420 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="922e33c0-945e-42eb-a0c6-b7a0262e9563" containerName="extract-utilities" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.221429 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="922e33c0-945e-42eb-a0c6-b7a0262e9563" containerName="extract-utilities" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.221726 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="922e33c0-945e-42eb-a0c6-b7a0262e9563" containerName="registry-server" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.222570 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.227643 4815 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.227953 4815 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.244425 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn"] Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.306375 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-secret-volume\") pod \"collect-profiles-29415525-78prn\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.306507 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-config-volume\") pod \"collect-profiles-29415525-78prn\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.306607 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wvhbj\" (UniqueName: \"kubernetes.io/projected/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-kube-api-access-wvhbj\") pod \"collect-profiles-29415525-78prn\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.408087 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-secret-volume\") pod \"collect-profiles-29415525-78prn\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.408167 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-config-volume\") pod \"collect-profiles-29415525-78prn\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.408232 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wvhbj\" (UniqueName: \"kubernetes.io/projected/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-kube-api-access-wvhbj\") pod \"collect-profiles-29415525-78prn\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.409738 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-config-volume\") pod \"collect-profiles-29415525-78prn\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.415923 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-secret-volume\") pod \"collect-profiles-29415525-78prn\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.433160 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wvhbj\" (UniqueName: \"kubernetes.io/projected/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-kube-api-access-wvhbj\") pod \"collect-profiles-29415525-78prn\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:00 crc kubenswrapper[4815]: I1205 10:45:00.555176 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:01 crc kubenswrapper[4815]: I1205 10:45:01.061878 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn"] Dec 05 10:45:01 crc kubenswrapper[4815]: I1205 10:45:01.560343 4815 generic.go:334] "Generic (PLEG): container finished" podID="7e37643f-ba67-49ce-b9c9-0ffe25ef4649" containerID="90ae7deff4086b406edaeb1d17c4cfecbac25c56f2badb89025721c307decc8a" exitCode=0 Dec 05 10:45:01 crc kubenswrapper[4815]: I1205 10:45:01.560426 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" event={"ID":"7e37643f-ba67-49ce-b9c9-0ffe25ef4649","Type":"ContainerDied","Data":"90ae7deff4086b406edaeb1d17c4cfecbac25c56f2badb89025721c307decc8a"} Dec 05 10:45:01 crc kubenswrapper[4815]: I1205 10:45:01.561667 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" event={"ID":"7e37643f-ba67-49ce-b9c9-0ffe25ef4649","Type":"ContainerStarted","Data":"a7c6fe2ebeda19d040a491fc70715e9fd3d6a75b7d5af53ef0a08b8c44c0345d"} Dec 05 10:45:02 crc kubenswrapper[4815]: I1205 10:45:02.938107 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:02 crc kubenswrapper[4815]: I1205 10:45:02.993747 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-secret-volume\") pod \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " Dec 05 10:45:02 crc kubenswrapper[4815]: I1205 10:45:02.993973 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wvhbj\" (UniqueName: \"kubernetes.io/projected/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-kube-api-access-wvhbj\") pod \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " Dec 05 10:45:02 crc kubenswrapper[4815]: I1205 10:45:02.994081 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-config-volume\") pod \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\" (UID: \"7e37643f-ba67-49ce-b9c9-0ffe25ef4649\") " Dec 05 10:45:02 crc kubenswrapper[4815]: I1205 10:45:02.996003 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-config-volume" (OuterVolumeSpecName: "config-volume") pod "7e37643f-ba67-49ce-b9c9-0ffe25ef4649" (UID: "7e37643f-ba67-49ce-b9c9-0ffe25ef4649"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 05 10:45:03 crc kubenswrapper[4815]: I1205 10:45:03.031228 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7e37643f-ba67-49ce-b9c9-0ffe25ef4649" (UID: "7e37643f-ba67-49ce-b9c9-0ffe25ef4649"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 05 10:45:03 crc kubenswrapper[4815]: I1205 10:45:03.041859 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-kube-api-access-wvhbj" (OuterVolumeSpecName: "kube-api-access-wvhbj") pod "7e37643f-ba67-49ce-b9c9-0ffe25ef4649" (UID: "7e37643f-ba67-49ce-b9c9-0ffe25ef4649"). InnerVolumeSpecName "kube-api-access-wvhbj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:45:03 crc kubenswrapper[4815]: I1205 10:45:03.097373 4815 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:45:03 crc kubenswrapper[4815]: I1205 10:45:03.097451 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wvhbj\" (UniqueName: \"kubernetes.io/projected/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-kube-api-access-wvhbj\") on node \"crc\" DevicePath \"\"" Dec 05 10:45:03 crc kubenswrapper[4815]: I1205 10:45:03.097463 4815 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7e37643f-ba67-49ce-b9c9-0ffe25ef4649-config-volume\") on node \"crc\" DevicePath \"\"" Dec 05 10:45:03 crc kubenswrapper[4815]: I1205 10:45:03.585160 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" event={"ID":"7e37643f-ba67-49ce-b9c9-0ffe25ef4649","Type":"ContainerDied","Data":"a7c6fe2ebeda19d040a491fc70715e9fd3d6a75b7d5af53ef0a08b8c44c0345d"} Dec 05 10:45:03 crc kubenswrapper[4815]: I1205 10:45:03.585234 4815 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7c6fe2ebeda19d040a491fc70715e9fd3d6a75b7d5af53ef0a08b8c44c0345d" Dec 05 10:45:03 crc kubenswrapper[4815]: I1205 10:45:03.585335 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29415525-78prn" Dec 05 10:45:04 crc kubenswrapper[4815]: I1205 10:45:04.030013 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb"] Dec 05 10:45:04 crc kubenswrapper[4815]: I1205 10:45:04.039313 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29415480-w77cb"] Dec 05 10:45:05 crc kubenswrapper[4815]: I1205 10:45:05.433288 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e42a353-a6f5-427a-bff2-64f6a9b80660" path="/var/lib/kubelet/pods/1e42a353-a6f5-427a-bff2-64f6a9b80660/volumes" Dec 05 10:45:08 crc kubenswrapper[4815]: I1205 10:45:08.420048 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:45:08 crc kubenswrapper[4815]: E1205 10:45:08.420932 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:45:11 crc kubenswrapper[4815]: I1205 10:45:11.052911 4815 scope.go:117] "RemoveContainer" containerID="57e2fa5a7af642c7f38603e6e3672abf5849358e0818b54d1851adae717aabfb" Dec 05 10:45:20 crc kubenswrapper[4815]: I1205 10:45:20.419339 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:45:20 crc kubenswrapper[4815]: E1205 10:45:20.420069 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:45:32 crc kubenswrapper[4815]: I1205 10:45:32.419587 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:45:32 crc kubenswrapper[4815]: E1205 10:45:32.420705 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:45:36 crc kubenswrapper[4815]: I1205 10:45:36.951049 4815 generic.go:334] "Generic (PLEG): container finished" podID="11a003bf-f178-4f1f-a45a-f13e24f6ce8c" containerID="fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be" exitCode=0 Dec 05 10:45:36 crc kubenswrapper[4815]: I1205 10:45:36.951182 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-lq64j/must-gather-ttd7k" event={"ID":"11a003bf-f178-4f1f-a45a-f13e24f6ce8c","Type":"ContainerDied","Data":"fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be"} Dec 05 10:45:36 crc kubenswrapper[4815]: I1205 10:45:36.952252 4815 scope.go:117] "RemoveContainer" containerID="fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be" Dec 05 10:45:37 crc kubenswrapper[4815]: I1205 10:45:37.449367 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lq64j_must-gather-ttd7k_11a003bf-f178-4f1f-a45a-f13e24f6ce8c/gather/0.log" Dec 05 10:45:43 crc kubenswrapper[4815]: I1205 10:45:43.418864 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:45:43 crc kubenswrapper[4815]: E1205 10:45:43.420126 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.061993 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-lq64j/must-gather-ttd7k"] Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.062744 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-lq64j/must-gather-ttd7k" podUID="11a003bf-f178-4f1f-a45a-f13e24f6ce8c" containerName="copy" containerID="cri-o://c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae" gracePeriod=2 Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.085397 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-lq64j/must-gather-ttd7k"] Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.489850 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lq64j_must-gather-ttd7k_11a003bf-f178-4f1f-a45a-f13e24f6ce8c/copy/0.log" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.490856 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/must-gather-ttd7k" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.562526 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cdc7\" (UniqueName: \"kubernetes.io/projected/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-kube-api-access-2cdc7\") pod \"11a003bf-f178-4f1f-a45a-f13e24f6ce8c\" (UID: \"11a003bf-f178-4f1f-a45a-f13e24f6ce8c\") " Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.562714 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-must-gather-output\") pod \"11a003bf-f178-4f1f-a45a-f13e24f6ce8c\" (UID: \"11a003bf-f178-4f1f-a45a-f13e24f6ce8c\") " Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.631268 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-kube-api-access-2cdc7" (OuterVolumeSpecName: "kube-api-access-2cdc7") pod "11a003bf-f178-4f1f-a45a-f13e24f6ce8c" (UID: "11a003bf-f178-4f1f-a45a-f13e24f6ce8c"). InnerVolumeSpecName "kube-api-access-2cdc7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.665431 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cdc7\" (UniqueName: \"kubernetes.io/projected/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-kube-api-access-2cdc7\") on node \"crc\" DevicePath \"\"" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.678990 4815 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-lq64j_must-gather-ttd7k_11a003bf-f178-4f1f-a45a-f13e24f6ce8c/copy/0.log" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.679392 4815 generic.go:334] "Generic (PLEG): container finished" podID="11a003bf-f178-4f1f-a45a-f13e24f6ce8c" containerID="c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae" exitCode=143 Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.679465 4815 scope.go:117] "RemoveContainer" containerID="c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.679482 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-lq64j/must-gather-ttd7k" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.709013 4815 scope.go:117] "RemoveContainer" containerID="fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.768910 4815 scope.go:117] "RemoveContainer" containerID="c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae" Dec 05 10:45:52 crc kubenswrapper[4815]: E1205 10:45:52.771980 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae\": container with ID starting with c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae not found: ID does not exist" containerID="c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.772021 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae"} err="failed to get container status \"c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae\": rpc error: code = NotFound desc = could not find container \"c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae\": container with ID starting with c628d6587d530552cd6bfd1844c3b5e5d8e39c620351a1c24023f83c33994dae not found: ID does not exist" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.772049 4815 scope.go:117] "RemoveContainer" containerID="fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be" Dec 05 10:45:52 crc kubenswrapper[4815]: E1205 10:45:52.772585 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be\": container with ID starting with fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be not found: ID does not exist" containerID="fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.772681 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be"} err="failed to get container status \"fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be\": rpc error: code = NotFound desc = could not find container \"fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be\": container with ID starting with fd115e3ba6acfae41024912626d315e80dc71966dc7d2b49b20569f274f290be not found: ID does not exist" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.828160 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "11a003bf-f178-4f1f-a45a-f13e24f6ce8c" (UID: "11a003bf-f178-4f1f-a45a-f13e24f6ce8c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:45:52 crc kubenswrapper[4815]: I1205 10:45:52.872835 4815 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/11a003bf-f178-4f1f-a45a-f13e24f6ce8c-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 05 10:45:53 crc kubenswrapper[4815]: I1205 10:45:53.428035 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11a003bf-f178-4f1f-a45a-f13e24f6ce8c" path="/var/lib/kubelet/pods/11a003bf-f178-4f1f-a45a-f13e24f6ce8c/volumes" Dec 05 10:45:55 crc kubenswrapper[4815]: I1205 10:45:55.419695 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:45:55 crc kubenswrapper[4815]: E1205 10:45:55.420432 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:46:09 crc kubenswrapper[4815]: I1205 10:46:09.420372 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:46:09 crc kubenswrapper[4815]: E1205 10:46:09.421243 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:46:23 crc kubenswrapper[4815]: I1205 10:46:23.419812 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:46:23 crc kubenswrapper[4815]: E1205 10:46:23.421218 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:46:37 crc kubenswrapper[4815]: I1205 10:46:37.419080 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:46:37 crc kubenswrapper[4815]: E1205 10:46:37.419956 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:46:50 crc kubenswrapper[4815]: I1205 10:46:50.418238 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:46:50 crc kubenswrapper[4815]: E1205 10:46:50.420103 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:47:03 crc kubenswrapper[4815]: I1205 10:47:03.418949 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:47:03 crc kubenswrapper[4815]: E1205 10:47:03.419874 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:47:17 crc kubenswrapper[4815]: I1205 10:47:17.419747 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:47:17 crc kubenswrapper[4815]: E1205 10:47:17.421133 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:47:29 crc kubenswrapper[4815]: I1205 10:47:29.419243 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:47:29 crc kubenswrapper[4815]: E1205 10:47:29.420052 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:47:44 crc kubenswrapper[4815]: I1205 10:47:44.418896 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:47:44 crc kubenswrapper[4815]: E1205 10:47:44.420117 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:47:58 crc kubenswrapper[4815]: I1205 10:47:58.418799 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:47:58 crc kubenswrapper[4815]: E1205 10:47:58.419808 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:48:11 crc kubenswrapper[4815]: I1205 10:48:11.423969 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:48:11 crc kubenswrapper[4815]: E1205 10:48:11.424655 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:48:26 crc kubenswrapper[4815]: I1205 10:48:26.419822 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:48:26 crc kubenswrapper[4815]: E1205 10:48:26.420885 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:48:41 crc kubenswrapper[4815]: I1205 10:48:41.427513 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:48:41 crc kubenswrapper[4815]: E1205 10:48:41.428339 4815 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-4wq7d_openshift-machine-config-operator(0a874c44-831d-4f81-bf48-2025f5615574)\"" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" podUID="0a874c44-831d-4f81-bf48-2025f5615574" Dec 05 10:48:55 crc kubenswrapper[4815]: I1205 10:48:55.419392 4815 scope.go:117] "RemoveContainer" containerID="4c33a0250d7e933ede20b348f1a61527c37f244f0b281c6b220c1d285b331a94" Dec 05 10:48:56 crc kubenswrapper[4815]: I1205 10:48:56.588246 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-4wq7d" event={"ID":"0a874c44-831d-4f81-bf48-2025f5615574","Type":"ContainerStarted","Data":"41d1f18d34eef7f7e8646f7fdd621b899ff3d64e3397f35ddac2a331dd5c4152"} Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.641892 4815 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-f272n"] Dec 05 10:49:07 crc kubenswrapper[4815]: E1205 10:49:07.642930 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11a003bf-f178-4f1f-a45a-f13e24f6ce8c" containerName="copy" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.642962 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="11a003bf-f178-4f1f-a45a-f13e24f6ce8c" containerName="copy" Dec 05 10:49:07 crc kubenswrapper[4815]: E1205 10:49:07.642996 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11a003bf-f178-4f1f-a45a-f13e24f6ce8c" containerName="gather" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.643004 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="11a003bf-f178-4f1f-a45a-f13e24f6ce8c" containerName="gather" Dec 05 10:49:07 crc kubenswrapper[4815]: E1205 10:49:07.643019 4815 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e37643f-ba67-49ce-b9c9-0ffe25ef4649" containerName="collect-profiles" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.643026 4815 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e37643f-ba67-49ce-b9c9-0ffe25ef4649" containerName="collect-profiles" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.643298 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e37643f-ba67-49ce-b9c9-0ffe25ef4649" containerName="collect-profiles" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.643317 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="11a003bf-f178-4f1f-a45a-f13e24f6ce8c" containerName="copy" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.643333 4815 memory_manager.go:354] "RemoveStaleState removing state" podUID="11a003bf-f178-4f1f-a45a-f13e24f6ce8c" containerName="gather" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.645110 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.658585 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f272n"] Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.815229 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfkb2\" (UniqueName: \"kubernetes.io/projected/8e92c010-7ead-4162-8f0f-d638280a4e55-kube-api-access-vfkb2\") pod \"redhat-marketplace-f272n\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.815731 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-utilities\") pod \"redhat-marketplace-f272n\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.815777 4815 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-catalog-content\") pod \"redhat-marketplace-f272n\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.917166 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-utilities\") pod \"redhat-marketplace-f272n\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.917225 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-catalog-content\") pod \"redhat-marketplace-f272n\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.917295 4815 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfkb2\" (UniqueName: \"kubernetes.io/projected/8e92c010-7ead-4162-8f0f-d638280a4e55-kube-api-access-vfkb2\") pod \"redhat-marketplace-f272n\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.917752 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-utilities\") pod \"redhat-marketplace-f272n\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.917812 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-catalog-content\") pod \"redhat-marketplace-f272n\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.952381 4815 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfkb2\" (UniqueName: \"kubernetes.io/projected/8e92c010-7ead-4162-8f0f-d638280a4e55-kube-api-access-vfkb2\") pod \"redhat-marketplace-f272n\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:07 crc kubenswrapper[4815]: I1205 10:49:07.962991 4815 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:08 crc kubenswrapper[4815]: I1205 10:49:08.487608 4815 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-f272n"] Dec 05 10:49:08 crc kubenswrapper[4815]: I1205 10:49:08.709626 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f272n" event={"ID":"8e92c010-7ead-4162-8f0f-d638280a4e55","Type":"ContainerStarted","Data":"6eb465e9c49fe8d4673d0cf14a92c39cafb78e5a4a6e03200abc8209aea1ee09"} Dec 05 10:49:08 crc kubenswrapper[4815]: E1205 10:49:08.943054 4815 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e92c010_7ead_4162_8f0f_d638280a4e55.slice/crio-conmon-f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e92c010_7ead_4162_8f0f_d638280a4e55.slice/crio-f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1.scope\": RecentStats: unable to find data in memory cache]" Dec 05 10:49:09 crc kubenswrapper[4815]: I1205 10:49:09.726278 4815 generic.go:334] "Generic (PLEG): container finished" podID="8e92c010-7ead-4162-8f0f-d638280a4e55" containerID="f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1" exitCode=0 Dec 05 10:49:09 crc kubenswrapper[4815]: I1205 10:49:09.726336 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f272n" event={"ID":"8e92c010-7ead-4162-8f0f-d638280a4e55","Type":"ContainerDied","Data":"f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1"} Dec 05 10:49:09 crc kubenswrapper[4815]: I1205 10:49:09.729412 4815 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 05 10:49:10 crc kubenswrapper[4815]: I1205 10:49:10.736511 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f272n" event={"ID":"8e92c010-7ead-4162-8f0f-d638280a4e55","Type":"ContainerStarted","Data":"818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122"} Dec 05 10:49:11 crc kubenswrapper[4815]: I1205 10:49:11.747070 4815 generic.go:334] "Generic (PLEG): container finished" podID="8e92c010-7ead-4162-8f0f-d638280a4e55" containerID="818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122" exitCode=0 Dec 05 10:49:11 crc kubenswrapper[4815]: I1205 10:49:11.747444 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f272n" event={"ID":"8e92c010-7ead-4162-8f0f-d638280a4e55","Type":"ContainerDied","Data":"818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122"} Dec 05 10:49:12 crc kubenswrapper[4815]: I1205 10:49:12.756600 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f272n" event={"ID":"8e92c010-7ead-4162-8f0f-d638280a4e55","Type":"ContainerStarted","Data":"ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7"} Dec 05 10:49:12 crc kubenswrapper[4815]: I1205 10:49:12.779663 4815 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-f272n" podStartSLOduration=3.321096743 podStartE2EDuration="5.779626465s" podCreationTimestamp="2025-12-05 10:49:07 +0000 UTC" firstStartedPulling="2025-12-05 10:49:09.729182838 +0000 UTC m=+6148.607789675" lastFinishedPulling="2025-12-05 10:49:12.18771256 +0000 UTC m=+6151.066319397" observedRunningTime="2025-12-05 10:49:12.777700433 +0000 UTC m=+6151.656307270" watchObservedRunningTime="2025-12-05 10:49:12.779626465 +0000 UTC m=+6151.658233322" Dec 05 10:49:17 crc kubenswrapper[4815]: I1205 10:49:17.964473 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:17 crc kubenswrapper[4815]: I1205 10:49:17.965078 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:18 crc kubenswrapper[4815]: I1205 10:49:18.022811 4815 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:18 crc kubenswrapper[4815]: I1205 10:49:18.876407 4815 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:18 crc kubenswrapper[4815]: I1205 10:49:18.931595 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f272n"] Dec 05 10:49:20 crc kubenswrapper[4815]: I1205 10:49:20.859325 4815 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-f272n" podUID="8e92c010-7ead-4162-8f0f-d638280a4e55" containerName="registry-server" containerID="cri-o://ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7" gracePeriod=2 Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.383836 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.458291 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfkb2\" (UniqueName: \"kubernetes.io/projected/8e92c010-7ead-4162-8f0f-d638280a4e55-kube-api-access-vfkb2\") pod \"8e92c010-7ead-4162-8f0f-d638280a4e55\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.459738 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-catalog-content\") pod \"8e92c010-7ead-4162-8f0f-d638280a4e55\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.460079 4815 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-utilities\") pod \"8e92c010-7ead-4162-8f0f-d638280a4e55\" (UID: \"8e92c010-7ead-4162-8f0f-d638280a4e55\") " Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.461115 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-utilities" (OuterVolumeSpecName: "utilities") pod "8e92c010-7ead-4162-8f0f-d638280a4e55" (UID: "8e92c010-7ead-4162-8f0f-d638280a4e55"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.462139 4815 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-utilities\") on node \"crc\" DevicePath \"\"" Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.465783 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e92c010-7ead-4162-8f0f-d638280a4e55-kube-api-access-vfkb2" (OuterVolumeSpecName: "kube-api-access-vfkb2") pod "8e92c010-7ead-4162-8f0f-d638280a4e55" (UID: "8e92c010-7ead-4162-8f0f-d638280a4e55"). InnerVolumeSpecName "kube-api-access-vfkb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.487532 4815 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8e92c010-7ead-4162-8f0f-d638280a4e55" (UID: "8e92c010-7ead-4162-8f0f-d638280a4e55"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.564107 4815 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8e92c010-7ead-4162-8f0f-d638280a4e55-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.564150 4815 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfkb2\" (UniqueName: \"kubernetes.io/projected/8e92c010-7ead-4162-8f0f-d638280a4e55-kube-api-access-vfkb2\") on node \"crc\" DevicePath \"\"" Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.948801 4815 generic.go:334] "Generic (PLEG): container finished" podID="8e92c010-7ead-4162-8f0f-d638280a4e55" containerID="ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7" exitCode=0 Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.949599 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f272n" event={"ID":"8e92c010-7ead-4162-8f0f-d638280a4e55","Type":"ContainerDied","Data":"ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7"} Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.949730 4815 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-f272n" event={"ID":"8e92c010-7ead-4162-8f0f-d638280a4e55","Type":"ContainerDied","Data":"6eb465e9c49fe8d4673d0cf14a92c39cafb78e5a4a6e03200abc8209aea1ee09"} Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.949808 4815 scope.go:117] "RemoveContainer" containerID="ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7" Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.950030 4815 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-f272n" Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.969277 4815 scope.go:117] "RemoveContainer" containerID="818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122" Dec 05 10:49:21 crc kubenswrapper[4815]: I1205 10:49:21.988835 4815 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-f272n"] Dec 05 10:49:22 crc kubenswrapper[4815]: I1205 10:49:22.008103 4815 scope.go:117] "RemoveContainer" containerID="f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1" Dec 05 10:49:22 crc kubenswrapper[4815]: I1205 10:49:22.060425 4815 scope.go:117] "RemoveContainer" containerID="ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7" Dec 05 10:49:22 crc kubenswrapper[4815]: E1205 10:49:22.061064 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7\": container with ID starting with ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7 not found: ID does not exist" containerID="ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7" Dec 05 10:49:22 crc kubenswrapper[4815]: I1205 10:49:22.061089 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7"} err="failed to get container status \"ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7\": rpc error: code = NotFound desc = could not find container \"ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7\": container with ID starting with ae6cb08319e6343796e46eea290c2da847cb838b351079cc0a44f4354fa66be7 not found: ID does not exist" Dec 05 10:49:22 crc kubenswrapper[4815]: I1205 10:49:22.061118 4815 scope.go:117] "RemoveContainer" containerID="818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122" Dec 05 10:49:22 crc kubenswrapper[4815]: E1205 10:49:22.061464 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122\": container with ID starting with 818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122 not found: ID does not exist" containerID="818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122" Dec 05 10:49:22 crc kubenswrapper[4815]: I1205 10:49:22.061502 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122"} err="failed to get container status \"818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122\": rpc error: code = NotFound desc = could not find container \"818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122\": container with ID starting with 818d881b44d96b14dd4c601cb222dad23d24c6ba75e4b59d0020eef9f21ee122 not found: ID does not exist" Dec 05 10:49:22 crc kubenswrapper[4815]: I1205 10:49:22.061531 4815 scope.go:117] "RemoveContainer" containerID="f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1" Dec 05 10:49:22 crc kubenswrapper[4815]: E1205 10:49:22.062938 4815 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1\": container with ID starting with f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1 not found: ID does not exist" containerID="f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1" Dec 05 10:49:22 crc kubenswrapper[4815]: I1205 10:49:22.062999 4815 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1"} err="failed to get container status \"f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1\": rpc error: code = NotFound desc = could not find container \"f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1\": container with ID starting with f2c34ee866aed52ad395f1713620dbd32807c9a90c613c6ae8cde33d9b1371b1 not found: ID does not exist" Dec 05 10:49:22 crc kubenswrapper[4815]: I1205 10:49:22.064635 4815 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-f272n"] Dec 05 10:49:23 crc kubenswrapper[4815]: I1205 10:49:23.434703 4815 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e92c010-7ead-4162-8f0f-d638280a4e55" path="/var/lib/kubelet/pods/8e92c010-7ead-4162-8f0f-d638280a4e55/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114534333024447 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114534334017365 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114517553016514 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114517554015465 5ustar corecore